var/home/core/zuul-output/0000755000175000017500000000000015066456370014541 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015066467204015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005250323715066467175017725 0ustar rootrootSep 29 10:22:07 crc systemd[1]: Starting Kubernetes Kubelet... Sep 29 10:22:07 crc restorecon[4675]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:07 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:08 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 10:22:09 crc restorecon[4675]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 29 10:22:10 crc kubenswrapper[4727]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.691801 4727 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699519 4727 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699579 4727 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699590 4727 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699599 4727 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699608 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699617 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699625 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699633 4727 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699642 4727 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699650 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699658 4727 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699667 4727 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699675 4727 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699684 4727 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699706 4727 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699715 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699723 4727 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699734 4727 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699745 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699754 4727 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699762 4727 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699771 4727 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699780 4727 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699788 4727 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699796 4727 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699804 4727 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699812 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699821 4727 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699828 4727 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699837 4727 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699846 4727 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699854 4727 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699862 4727 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699870 4727 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699877 4727 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699888 4727 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699897 4727 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699948 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699957 4727 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699965 4727 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699973 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699981 4727 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699990 4727 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.699998 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700006 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700014 4727 feature_gate.go:330] unrecognized feature gate: Example Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700023 4727 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700031 4727 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700039 4727 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700048 4727 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700057 4727 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700064 4727 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700072 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700080 4727 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700087 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700098 4727 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700108 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700118 4727 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700128 4727 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700137 4727 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700144 4727 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700152 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700160 4727 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700170 4727 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700177 4727 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700185 4727 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700192 4727 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700200 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700208 4727 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700216 4727 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.700224 4727 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701129 4727 flags.go:64] FLAG: --address="0.0.0.0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701152 4727 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701171 4727 flags.go:64] FLAG: --anonymous-auth="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701184 4727 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701196 4727 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701206 4727 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701219 4727 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701230 4727 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701240 4727 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701250 4727 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701260 4727 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701270 4727 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701280 4727 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701290 4727 flags.go:64] FLAG: --cgroup-root="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701298 4727 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701308 4727 flags.go:64] FLAG: --client-ca-file="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701316 4727 flags.go:64] FLAG: --cloud-config="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701326 4727 flags.go:64] FLAG: --cloud-provider="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701361 4727 flags.go:64] FLAG: --cluster-dns="[]" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701373 4727 flags.go:64] FLAG: --cluster-domain="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701382 4727 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701392 4727 flags.go:64] FLAG: --config-dir="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701401 4727 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701411 4727 flags.go:64] FLAG: --container-log-max-files="5" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701424 4727 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701433 4727 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701443 4727 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701452 4727 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701462 4727 flags.go:64] FLAG: --contention-profiling="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701471 4727 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701480 4727 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701490 4727 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701500 4727 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701511 4727 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701521 4727 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701530 4727 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701539 4727 flags.go:64] FLAG: --enable-load-reader="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701548 4727 flags.go:64] FLAG: --enable-server="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701559 4727 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701572 4727 flags.go:64] FLAG: --event-burst="100" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701582 4727 flags.go:64] FLAG: --event-qps="50" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701591 4727 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701600 4727 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701609 4727 flags.go:64] FLAG: --eviction-hard="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701620 4727 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701631 4727 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701640 4727 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701649 4727 flags.go:64] FLAG: --eviction-soft="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701658 4727 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701667 4727 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701677 4727 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701686 4727 flags.go:64] FLAG: --experimental-mounter-path="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701695 4727 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701704 4727 flags.go:64] FLAG: --fail-swap-on="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701714 4727 flags.go:64] FLAG: --feature-gates="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701725 4727 flags.go:64] FLAG: --file-check-frequency="20s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701735 4727 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701744 4727 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701754 4727 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701763 4727 flags.go:64] FLAG: --healthz-port="10248" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701773 4727 flags.go:64] FLAG: --help="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701783 4727 flags.go:64] FLAG: --hostname-override="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701792 4727 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701801 4727 flags.go:64] FLAG: --http-check-frequency="20s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701811 4727 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701821 4727 flags.go:64] FLAG: --image-credential-provider-config="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701830 4727 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701839 4727 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701848 4727 flags.go:64] FLAG: --image-service-endpoint="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701857 4727 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701868 4727 flags.go:64] FLAG: --kube-api-burst="100" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701878 4727 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701888 4727 flags.go:64] FLAG: --kube-api-qps="50" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701896 4727 flags.go:64] FLAG: --kube-reserved="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701906 4727 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701916 4727 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701926 4727 flags.go:64] FLAG: --kubelet-cgroups="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701935 4727 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701944 4727 flags.go:64] FLAG: --lock-file="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701953 4727 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701962 4727 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701972 4727 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701985 4727 flags.go:64] FLAG: --log-json-split-stream="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.701995 4727 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702004 4727 flags.go:64] FLAG: --log-text-split-stream="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702013 4727 flags.go:64] FLAG: --logging-format="text" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702022 4727 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702032 4727 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702041 4727 flags.go:64] FLAG: --manifest-url="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702050 4727 flags.go:64] FLAG: --manifest-url-header="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702062 4727 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702072 4727 flags.go:64] FLAG: --max-open-files="1000000" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702083 4727 flags.go:64] FLAG: --max-pods="110" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702093 4727 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702102 4727 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702111 4727 flags.go:64] FLAG: --memory-manager-policy="None" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702121 4727 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702131 4727 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702140 4727 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702149 4727 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702169 4727 flags.go:64] FLAG: --node-status-max-images="50" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702179 4727 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702189 4727 flags.go:64] FLAG: --oom-score-adj="-999" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702198 4727 flags.go:64] FLAG: --pod-cidr="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702207 4727 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702221 4727 flags.go:64] FLAG: --pod-manifest-path="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702230 4727 flags.go:64] FLAG: --pod-max-pids="-1" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702240 4727 flags.go:64] FLAG: --pods-per-core="0" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702249 4727 flags.go:64] FLAG: --port="10250" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702258 4727 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702269 4727 flags.go:64] FLAG: --provider-id="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702278 4727 flags.go:64] FLAG: --qos-reserved="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702287 4727 flags.go:64] FLAG: --read-only-port="10255" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702296 4727 flags.go:64] FLAG: --register-node="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702305 4727 flags.go:64] FLAG: --register-schedulable="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702315 4727 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702330 4727 flags.go:64] FLAG: --registry-burst="10" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702361 4727 flags.go:64] FLAG: --registry-qps="5" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702370 4727 flags.go:64] FLAG: --reserved-cpus="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702379 4727 flags.go:64] FLAG: --reserved-memory="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702397 4727 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702407 4727 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702416 4727 flags.go:64] FLAG: --rotate-certificates="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702425 4727 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702434 4727 flags.go:64] FLAG: --runonce="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702443 4727 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702453 4727 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702463 4727 flags.go:64] FLAG: --seccomp-default="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702472 4727 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702481 4727 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702490 4727 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702500 4727 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702509 4727 flags.go:64] FLAG: --storage-driver-password="root" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702519 4727 flags.go:64] FLAG: --storage-driver-secure="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702528 4727 flags.go:64] FLAG: --storage-driver-table="stats" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702538 4727 flags.go:64] FLAG: --storage-driver-user="root" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702547 4727 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702557 4727 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702566 4727 flags.go:64] FLAG: --system-cgroups="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702575 4727 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702590 4727 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702599 4727 flags.go:64] FLAG: --tls-cert-file="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702608 4727 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702619 4727 flags.go:64] FLAG: --tls-min-version="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702628 4727 flags.go:64] FLAG: --tls-private-key-file="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702637 4727 flags.go:64] FLAG: --topology-manager-policy="none" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702647 4727 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702656 4727 flags.go:64] FLAG: --topology-manager-scope="container" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702666 4727 flags.go:64] FLAG: --v="2" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702677 4727 flags.go:64] FLAG: --version="false" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702689 4727 flags.go:64] FLAG: --vmodule="" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702700 4727 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.702713 4727 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702930 4727 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702941 4727 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702951 4727 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702961 4727 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702969 4727 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702978 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702986 4727 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.702994 4727 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703003 4727 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703011 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703018 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703027 4727 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703035 4727 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703042 4727 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703083 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703093 4727 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703104 4727 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703113 4727 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703174 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703187 4727 feature_gate.go:330] unrecognized feature gate: Example Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703198 4727 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703209 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703221 4727 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703235 4727 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703248 4727 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703260 4727 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703269 4727 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703280 4727 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703291 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703303 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703314 4727 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703329 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703369 4727 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703380 4727 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703390 4727 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703400 4727 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703409 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703419 4727 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703429 4727 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703440 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703450 4727 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703460 4727 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703469 4727 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703477 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703485 4727 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703493 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703506 4727 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703513 4727 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703521 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703529 4727 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703537 4727 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703544 4727 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703552 4727 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703561 4727 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703569 4727 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703576 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703584 4727 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703592 4727 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703600 4727 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703608 4727 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703616 4727 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703626 4727 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703637 4727 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703649 4727 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703657 4727 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703667 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703675 4727 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703685 4727 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703694 4727 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703703 4727 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.703713 4727 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.703726 4727 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.719585 4727 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.719642 4727 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719802 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719840 4727 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719852 4727 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719863 4727 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719872 4727 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719880 4727 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719944 4727 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719969 4727 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719985 4727 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.719996 4727 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720007 4727 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720018 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720028 4727 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720040 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720055 4727 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720072 4727 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720082 4727 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720091 4727 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720101 4727 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720111 4727 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720121 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720131 4727 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720141 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720152 4727 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720161 4727 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720171 4727 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720181 4727 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720192 4727 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720203 4727 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720213 4727 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720223 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720233 4727 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720243 4727 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720254 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720264 4727 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720274 4727 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720283 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720293 4727 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720302 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720313 4727 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720322 4727 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720332 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720389 4727 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720404 4727 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720415 4727 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720425 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720435 4727 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720444 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720455 4727 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720465 4727 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720473 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720481 4727 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720490 4727 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720498 4727 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720506 4727 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720514 4727 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720521 4727 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720530 4727 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720540 4727 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720548 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720556 4727 feature_gate.go:330] unrecognized feature gate: Example Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720564 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720572 4727 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720580 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720588 4727 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720596 4727 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720603 4727 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720611 4727 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720622 4727 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720632 4727 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720644 4727 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.720659 4727 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720925 4727 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720960 4727 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720983 4727 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.720996 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721005 4727 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721013 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721024 4727 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721032 4727 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721040 4727 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721047 4727 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721055 4727 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721063 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721072 4727 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721080 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721088 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721096 4727 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721104 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721111 4727 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721119 4727 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721127 4727 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721134 4727 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721142 4727 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721150 4727 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721157 4727 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721166 4727 feature_gate.go:330] unrecognized feature gate: Example Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721173 4727 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721181 4727 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721189 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721196 4727 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721205 4727 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721212 4727 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721223 4727 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721233 4727 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721241 4727 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721305 4727 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721315 4727 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721323 4727 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721379 4727 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721402 4727 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721413 4727 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721424 4727 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721433 4727 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721443 4727 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721454 4727 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721463 4727 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721473 4727 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721481 4727 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721489 4727 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721497 4727 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721505 4727 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721513 4727 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721520 4727 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721528 4727 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721535 4727 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721543 4727 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721551 4727 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721558 4727 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721566 4727 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721576 4727 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721586 4727 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721595 4727 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721603 4727 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721611 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721619 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721628 4727 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721638 4727 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721649 4727 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721658 4727 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721667 4727 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721675 4727 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 10:22:10 crc kubenswrapper[4727]: W0929 10:22:10.721685 4727 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.721698 4727 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.723101 4727 server.go:940] "Client rotation is on, will bootstrap in background" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.729298 4727 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.729468 4727 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.731604 4727 server.go:997] "Starting client certificate rotation" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.731661 4727 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.733319 4727 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-07 07:42:45.866987691 +0000 UTC Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.733462 4727 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2397h20m35.133530709s for next certificate rotation Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.854970 4727 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.861098 4727 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.898481 4727 log.go:25] "Validated CRI v1 runtime API" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.974282 4727 log.go:25] "Validated CRI v1 image API" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.977778 4727 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.985278 4727 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-29-10-17-23-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 29 10:22:10 crc kubenswrapper[4727]: I0929 10:22:10.985314 4727 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:10.999898 4727 manager.go:217] Machine: {Timestamp:2025-09-29 10:22:10.997655081 +0000 UTC m=+1.170968463 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:833ae235-acd8-410a-9184-a9bb33bace31 BootID:a4ffff91-59e9-4746-8297-f21cac254b29 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b3:2b:9a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b3:2b:9a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:69:25:42 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ef:4f:7f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1c:0b:5b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:53:7f:00 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:4b:39:ba:b3:df Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ae:a1:f2:9b:82:0d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.000185 4727 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.000325 4727 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.003860 4727 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.004131 4727 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.004180 4727 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.004507 4727 topology_manager.go:138] "Creating topology manager with none policy" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.004522 4727 container_manager_linux.go:303] "Creating device plugin manager" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.005147 4727 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.005192 4727 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.005588 4727 state_mem.go:36] "Initialized new in-memory state store" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.005701 4727 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.013207 4727 kubelet.go:418] "Attempting to sync node with API server" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.013237 4727 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.013260 4727 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.013273 4727 kubelet.go:324] "Adding apiserver pod source" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.013287 4727 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.019620 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.019717 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.020695 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.020755 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.025829 4727 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.027204 4727 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.032070 4727 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034904 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034939 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034950 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034961 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034980 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.034992 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035005 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035024 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035038 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035048 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035061 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.035070 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.037626 4727 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.038187 4727 server.go:1280] "Started kubelet" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.038433 4727 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.038540 4727 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.039656 4727 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.040073 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc systemd[1]: Started Kubernetes Kubelet. Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042180 4727 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042231 4727 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042271 4727 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 05:35:07.769261601 +0000 UTC Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042381 4727 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2659h12m56.726887838s for next certificate rotation Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042526 4727 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042543 4727 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.042514 4727 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.042575 4727 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.043109 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.043180 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.043597 4727 factory.go:55] Registering systemd factory Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.044191 4727 server.go:460] "Adding debug handlers to kubelet server" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.044512 4727 factory.go:221] Registration of the systemd container factory successfully Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.045233 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="200ms" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.045323 4727 factory.go:153] Registering CRI-O factory Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.045366 4727 factory.go:221] Registration of the crio container factory successfully Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.045425 4727 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.045450 4727 factory.go:103] Registering Raw factory Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.045469 4727 manager.go:1196] Started watching for new ooms in manager Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.046536 4727 manager.go:319] Starting recovery of all containers Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.046255 4727 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1869b9b8623c9bae default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-29 10:22:11.03814955 +0000 UTC m=+1.211462922,LastTimestamp:2025-09-29 10:22:11.03814955 +0000 UTC m=+1.211462922,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063439 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063512 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063527 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063537 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063549 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063559 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063571 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063585 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063600 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063613 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063625 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063635 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063646 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063659 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063673 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063682 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063692 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.063702 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066124 4727 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066156 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066173 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066186 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066218 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066228 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066238 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066249 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066260 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066271 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066281 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066292 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066302 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066310 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066320 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066358 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066393 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066403 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066416 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066426 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066436 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066446 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066456 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066467 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066477 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066488 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066500 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066509 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066520 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066532 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066541 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066550 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066559 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066568 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066578 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066590 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066600 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066612 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066624 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066635 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066646 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066656 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066667 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066677 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066688 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066698 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066710 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066721 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066733 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066745 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066787 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066802 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066814 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066825 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066837 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066848 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066860 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066872 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066885 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066896 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066908 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066920 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066932 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066945 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066962 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066974 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.066990 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067004 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067015 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067028 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067039 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067052 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067065 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067077 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067091 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067104 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067115 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067129 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067148 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067162 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067178 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067191 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067211 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067222 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067263 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067276 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067289 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067310 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067363 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067381 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067401 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067415 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067429 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067441 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067456 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067471 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067483 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067497 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067512 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067526 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067538 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067550 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067565 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067577 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067590 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067604 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067617 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067629 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067647 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067659 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067672 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067685 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067700 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067714 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067767 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067782 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067794 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067807 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067819 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067831 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067844 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067857 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067869 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067881 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067893 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067904 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067914 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067928 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067941 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067953 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067966 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067978 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.067990 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068002 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068013 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068024 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068035 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068045 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068057 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068068 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068107 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068118 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068129 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068140 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068152 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068164 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068175 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068187 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068199 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068211 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068223 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068235 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068247 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068257 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068268 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068279 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068289 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068301 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068314 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068329 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068356 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068368 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068382 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068396 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068409 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068423 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068435 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068448 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068461 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068475 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068488 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068506 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068519 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068533 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068546 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068559 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068573 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068587 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068602 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068614 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068626 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068639 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068656 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068669 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068682 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068694 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068707 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068721 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068734 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068746 4727 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068758 4727 reconstruct.go:97] "Volume reconstruction finished" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.068768 4727 reconciler.go:26] "Reconciler: start to sync state" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.073570 4727 manager.go:324] Recovery completed Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.083785 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.085291 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.085327 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.085365 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.087911 4727 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.087928 4727 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.087950 4727 state_mem.go:36] "Initialized new in-memory state store" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.104187 4727 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.107030 4727 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.107079 4727 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.107114 4727 kubelet.go:2335] "Starting kubelet main sync loop" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.107165 4727 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.107907 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.107974 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.125609 4727 policy_none.go:49] "None policy: Start" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.128363 4727 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.128403 4727 state_mem.go:35] "Initializing new in-memory state store" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.142656 4727 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.186589 4727 manager.go:334] "Starting Device Plugin manager" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.186671 4727 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.186692 4727 server.go:79] "Starting device plugin registration server" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.187265 4727 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.187299 4727 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.193281 4727 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.207459 4727 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.207559 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.208703 4727 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.208842 4727 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.208856 4727 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209122 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209175 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209388 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209460 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.209501 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210292 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210303 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210347 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210363 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210376 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210460 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210561 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.210593 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211050 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211081 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211247 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211394 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211527 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211568 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211568 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.211976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212016 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212137 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212469 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212504 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212697 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212710 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212811 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212836 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212844 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.212869 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.213412 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.213438 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.213451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.214211 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.214429 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.214615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.246238 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="400ms" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.270966 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271012 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271036 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271081 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271102 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271177 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271241 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271266 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271286 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271307 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271380 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271446 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271521 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.271566 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.288619 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.290098 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.290151 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.290168 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.290203 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.290812 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373186 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373255 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373283 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373309 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373391 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373426 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373450 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373543 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373581 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373545 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373635 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373650 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373721 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373606 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373726 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373599 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373801 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373746 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373824 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373832 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373850 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373855 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373871 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373873 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373913 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373925 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373941 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373950 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373987 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.373968 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.491745 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.493421 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.493484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.493495 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.493537 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.494230 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.548256 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.568101 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.593887 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.607889 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.615705 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.647256 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="800ms" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.651231 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-b2f2dd06c9f74ca4dfb887db623c32faec353769a08249446983fbc7f9bc6c0c WatchSource:0}: Error finding container b2f2dd06c9f74ca4dfb887db623c32faec353769a08249446983fbc7f9bc6c0c: Status 404 returned error can't find the container with id b2f2dd06c9f74ca4dfb887db623c32faec353769a08249446983fbc7f9bc6c0c Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.651984 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7c14c02d637b1e3a010771e343cca5a95ac3f0128fb566c47a3467ee2da7ef45 WatchSource:0}: Error finding container 7c14c02d637b1e3a010771e343cca5a95ac3f0128fb566c47a3467ee2da7ef45: Status 404 returned error can't find the container with id 7c14c02d637b1e3a010771e343cca5a95ac3f0128fb566c47a3467ee2da7ef45 Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.659580 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7d945ea7610e16e5924324f6e408236b9c86f45b6b44285e8798525bcdbd91bf WatchSource:0}: Error finding container 7d945ea7610e16e5924324f6e408236b9c86f45b6b44285e8798525bcdbd91bf: Status 404 returned error can't find the container with id 7d945ea7610e16e5924324f6e408236b9c86f45b6b44285e8798525bcdbd91bf Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.660403 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c4a1f5a0e83a5c5c40cde10ffbc0b014d982a9365b4b83f40cd8842f60360829 WatchSource:0}: Error finding container c4a1f5a0e83a5c5c40cde10ffbc0b014d982a9365b4b83f40cd8842f60360829: Status 404 returned error can't find the container with id c4a1f5a0e83a5c5c40cde10ffbc0b014d982a9365b4b83f40cd8842f60360829 Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.663134 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6810ad9dad18bed9996b144e5e4ffaf0f0ded09866b3cbea485c520d393ce4e9 WatchSource:0}: Error finding container 6810ad9dad18bed9996b144e5e4ffaf0f0ded09866b3cbea485c520d393ce4e9: Status 404 returned error can't find the container with id 6810ad9dad18bed9996b144e5e4ffaf0f0ded09866b3cbea485c520d393ce4e9 Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.895172 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.896466 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.896499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.896512 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:11 crc kubenswrapper[4727]: I0929 10:22:11.896538 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.896995 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:11 crc kubenswrapper[4727]: W0929 10:22:11.976325 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:11 crc kubenswrapper[4727]: E0929 10:22:11.976463 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.045038 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:12 crc kubenswrapper[4727]: W0929 10:22:12.055317 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:12 crc kubenswrapper[4727]: E0929 10:22:12.055513 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.118651 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b2f2dd06c9f74ca4dfb887db623c32faec353769a08249446983fbc7f9bc6c0c"} Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.120383 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6810ad9dad18bed9996b144e5e4ffaf0f0ded09866b3cbea485c520d393ce4e9"} Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.121455 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c4a1f5a0e83a5c5c40cde10ffbc0b014d982a9365b4b83f40cd8842f60360829"} Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.122307 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d945ea7610e16e5924324f6e408236b9c86f45b6b44285e8798525bcdbd91bf"} Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.123416 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7c14c02d637b1e3a010771e343cca5a95ac3f0128fb566c47a3467ee2da7ef45"} Sep 29 10:22:12 crc kubenswrapper[4727]: E0929 10:22:12.448830 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="1.6s" Sep 29 10:22:12 crc kubenswrapper[4727]: W0929 10:22:12.534811 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:12 crc kubenswrapper[4727]: E0929 10:22:12.535017 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:12 crc kubenswrapper[4727]: W0929 10:22:12.534839 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:12 crc kubenswrapper[4727]: E0929 10:22:12.535202 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.698142 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.700123 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.700180 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.700195 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:12 crc kubenswrapper[4727]: I0929 10:22:12.700228 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:12 crc kubenswrapper[4727]: E0929 10:22:12.700884 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:13 crc kubenswrapper[4727]: I0929 10:22:13.041385 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.041301 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:14 crc kubenswrapper[4727]: E0929 10:22:14.050106 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="3.2s" Sep 29 10:22:14 crc kubenswrapper[4727]: W0929 10:22:14.184701 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:14 crc kubenswrapper[4727]: E0929 10:22:14.185484 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.301871 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.304728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.304799 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.304821 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:14 crc kubenswrapper[4727]: I0929 10:22:14.304865 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:14 crc kubenswrapper[4727]: E0929 10:22:14.305620 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:14 crc kubenswrapper[4727]: W0929 10:22:14.492950 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:14 crc kubenswrapper[4727]: E0929 10:22:14.493059 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:14 crc kubenswrapper[4727]: W0929 10:22:14.827070 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:14 crc kubenswrapper[4727]: E0929 10:22:14.827202 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.041599 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.135432 4727 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494" exitCode=0 Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.135516 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.135546 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.136575 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.136616 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.136628 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138221 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138728 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138792 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.138998 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.140964 4727 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d" exitCode=0 Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.141064 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.141063 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.142105 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.142131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.142142 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.180918 4727 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14" exitCode=0 Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.181062 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.181112 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.182421 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.182464 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.182478 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.183255 4727 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a" exitCode=0 Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.183324 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a"} Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.183426 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.185538 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.185584 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:15 crc kubenswrapper[4727]: I0929 10:22:15.185594 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:15 crc kubenswrapper[4727]: W0929 10:22:15.545096 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:15 crc kubenswrapper[4727]: E0929 10:22:15.545203 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.042043 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.195300 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.195373 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.195385 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.199646 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.199718 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.199793 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.200632 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.200664 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.200675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.203529 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.203588 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.203600 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.203631 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.208194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.208237 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.208248 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.209808 4727 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce" exitCode=0 Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.209891 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.209923 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.211087 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.211123 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.211133 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.213755 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1"} Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.213821 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.215071 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.215096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.215107 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.268481 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.268975 4727 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": dial tcp 192.168.126.11:10357: connect: connection refused" start-of-body= Sep 29 10:22:16 crc kubenswrapper[4727]: I0929 10:22:16.269047 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": dial tcp 192.168.126.11:10357: connect: connection refused" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.041594 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.221387 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa"} Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.221457 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f"} Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224243 4727 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336" exitCode=0 Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224425 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224442 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224475 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224429 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224491 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336"} Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.224429 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226053 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226055 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226153 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226155 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226473 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.226526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.227005 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.227025 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.227034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:17 crc kubenswrapper[4727]: E0929 10:22:17.251732 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="6.4s" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.506149 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.507947 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.507993 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.508004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:17 crc kubenswrapper[4727]: I0929 10:22:17.508031 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:17 crc kubenswrapper[4727]: E0929 10:22:17.508530 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.040914 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.228313 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.229110 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.233937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.233994 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.234007 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.235459 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.235552 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:18 crc kubenswrapper[4727]: I0929 10:22:18.235581 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:18 crc kubenswrapper[4727]: W0929 10:22:18.343374 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:18 crc kubenswrapper[4727]: E0929 10:22:18.343515 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.041184 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.236882 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a"} Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.502169 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.502398 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.503667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.503760 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.503782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:19 crc kubenswrapper[4727]: W0929 10:22:19.543650 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:19 crc kubenswrapper[4727]: E0929 10:22:19.543796 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.688655 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:19 crc kubenswrapper[4727]: W0929 10:22:19.728132 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:19 crc kubenswrapper[4727]: E0929 10:22:19.728208 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.729509 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:19 crc kubenswrapper[4727]: I0929 10:22:19.952407 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.041492 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.244175 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43"} Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.244248 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd"} Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.244263 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39"} Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.244276 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.245170 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.245202 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.245213 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.682295 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.682836 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.685116 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.685179 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:20 crc kubenswrapper[4727]: I0929 10:22:20.685205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.176879 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.177099 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.177402 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.178225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.178278 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.178293 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:21 crc kubenswrapper[4727]: E0929 10:22:21.193426 4727 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.252215 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a"} Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.252326 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.252371 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.252406 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.254994 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255523 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255578 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255575 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255664 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.255684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.609959 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:21 crc kubenswrapper[4727]: I0929 10:22:21.696514 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.254637 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.254726 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.255697 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.255731 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.255741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.255925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.255990 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:22 crc kubenswrapper[4727]: I0929 10:22:22.256009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.258903 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.259941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.259977 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.259989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.909103 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.910921 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.910959 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.910973 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:23 crc kubenswrapper[4727]: I0929 10:22:23.910999 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:28 crc kubenswrapper[4727]: I0929 10:22:28.937954 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 29 10:22:28 crc kubenswrapper[4727]: I0929 10:22:28.938403 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:28 crc kubenswrapper[4727]: I0929 10:22:28.939791 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:28 crc kubenswrapper[4727]: I0929 10:22:28.939825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:28 crc kubenswrapper[4727]: I0929 10:22:28.939836 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.268710 4727 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.268844 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.959728 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.959934 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.961927 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.962033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:29 crc kubenswrapper[4727]: I0929 10:22:29.962052 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:30 crc kubenswrapper[4727]: E0929 10:22:30.942318 4727 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.1869b9b8623c9bae default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-29 10:22:11.03814955 +0000 UTC m=+1.211462922,LastTimestamp:2025-09-29 10:22:11.03814955 +0000 UTC m=+1.211462922,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.041287 4727 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.177168 4727 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.177245 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 10:22:31 crc kubenswrapper[4727]: E0929 10:22:31.193639 4727 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.281372 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.284098 4727 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa" exitCode=255 Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.284165 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa"} Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.284415 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.285450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.286433 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.286445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.287073 4727 scope.go:117] "RemoveContainer" containerID="d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa" Sep 29 10:22:31 crc kubenswrapper[4727]: W0929 10:22:31.371044 4727 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.371203 4727 trace.go:236] Trace[652296989]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 10:22:21.369) (total time: 10001ms): Sep 29 10:22:31 crc kubenswrapper[4727]: Trace[652296989]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:22:31.371) Sep 29 10:22:31 crc kubenswrapper[4727]: Trace[652296989]: [10.001240968s] [10.001240968s] END Sep 29 10:22:31 crc kubenswrapper[4727]: E0929 10:22:31.371242 4727 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.657553 4727 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 29 10:22:31 crc kubenswrapper[4727]: I0929 10:22:31.657617 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.289607 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.292643 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e"} Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.292822 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.293838 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.293890 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:32 crc kubenswrapper[4727]: I0929 10:22:32.293906 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.183600 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.183820 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.184471 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.185537 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.185587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.185602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.188990 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.303886 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.305386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.305439 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.305450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:36 crc kubenswrapper[4727]: E0929 10:22:36.651803 4727 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.654972 4727 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 29 10:22:36 crc kubenswrapper[4727]: E0929 10:22:36.660003 4727 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.661919 4727 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.661979 4727 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.666969 4727 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.709507 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:36 crc kubenswrapper[4727]: I0929 10:22:36.715268 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.028114 4727 apiserver.go:52] "Watching apiserver" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.031768 4727 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.031952 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.032306 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.032449 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.032513 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.032556 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.032860 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.033140 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.033191 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.033217 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.033252 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.035610 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.035627 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.035752 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.036264 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.036307 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.036376 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.036625 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.037238 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.041249 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.043191 4727 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064075 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064150 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064187 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064221 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064254 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064277 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064302 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064324 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064386 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064415 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064490 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064791 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064890 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064907 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064952 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064960 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064968 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.064989 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065015 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065194 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065197 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065232 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065258 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065271 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065276 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065361 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065316 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065393 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065521 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065595 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065618 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065641 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065660 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065697 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065730 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065813 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065876 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065923 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065878 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065877 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.065958 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066013 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066053 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066068 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066092 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066104 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066141 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066170 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066187 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066201 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066236 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066244 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066266 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066298 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066330 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066377 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066410 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066447 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066503 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066507 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066548 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066584 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066614 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066627 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066641 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066709 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066776 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066782 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.066975 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067011 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067016 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067068 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067101 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067128 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067153 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067181 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067211 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067237 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067262 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067288 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067313 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067354 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067380 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067404 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067460 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067488 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067514 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067538 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067565 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067589 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067615 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067638 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067661 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067685 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067710 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067740 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067768 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067793 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067815 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067838 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067860 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067884 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067912 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067935 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067957 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067983 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068009 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068036 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068062 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068087 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068115 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068140 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068164 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068189 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068220 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068246 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068270 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068293 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068326 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068368 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068396 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068419 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068457 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068479 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068506 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068530 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068554 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068574 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068594 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068615 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068636 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068658 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068685 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068705 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068727 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068748 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068770 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068792 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068817 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068838 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068862 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068883 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068904 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068924 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068946 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068969 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068995 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069020 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069043 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069063 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069085 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069110 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069135 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069157 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069180 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069202 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069225 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069250 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069271 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069295 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069317 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067060 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067168 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067233 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067269 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067391 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067542 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067619 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069586 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070173 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070213 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070242 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070272 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070298 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070324 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070427 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070455 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070477 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072021 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072061 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072091 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072117 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072145 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072169 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072193 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072217 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072242 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072267 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072292 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072313 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072352 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072380 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072405 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072429 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072457 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072482 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072506 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072528 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072553 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072577 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072604 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072627 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072648 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072665 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072682 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072698 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072715 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072731 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072747 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072774 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072794 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072810 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072827 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072845 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072864 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072885 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072906 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072927 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072947 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072967 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072991 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073013 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073036 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073057 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073081 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073103 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073127 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073182 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073211 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073237 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073261 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073285 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073313 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073368 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073395 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073417 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073443 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073470 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073494 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073522 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073548 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073631 4727 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073647 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073659 4727 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073673 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073684 4727 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073696 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073710 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073722 4727 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073734 4727 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073751 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073764 4727 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073776 4727 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073789 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073801 4727 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073814 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073827 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073839 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073852 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073865 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073877 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073890 4727 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073903 4727 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073915 4727 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073928 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073941 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073955 4727 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073969 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073982 4727 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073997 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074011 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074024 4727 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074036 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074769 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067698 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067732 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067881 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067963 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.067983 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068012 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068030 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068108 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068208 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068227 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068621 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068961 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.068980 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069106 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069111 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069249 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069914 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070082 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.069332 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070202 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070283 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.070456 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071020 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071265 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071360 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071600 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071619 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.071616 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072125 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.072501 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073240 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073407 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073464 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.073913 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074075 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074394 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074601 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074783 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.074973 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075103 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075275 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075354 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075479 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075485 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082554 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075263 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075818 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082586 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075877 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.075926 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076117 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076137 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076159 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076258 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076295 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076486 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076535 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076550 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.076709 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.077863 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078038 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078020 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078329 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078663 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078691 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078769 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078944 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.079202 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.079602 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.079672 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.079736 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.078737 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.079860 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080178 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080182 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080464 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080711 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080733 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080740 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.080903 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.081816 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.081839 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.081891 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082685 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082890 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082943 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.083258 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.083528 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.083823 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.083839 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.084127 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.084266 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.084414 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.084447 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.084619 4727 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.085587 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.082892 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.085637 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.085833 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.086122 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.086147 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.086430 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.086673 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.086964 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.086967 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.087183 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.087406 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.088623 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.087672 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.087793 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:37.587763529 +0000 UTC m=+27.761077071 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.088164 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.088573 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.088874 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.088916 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.089016 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:37.58893733 +0000 UTC m=+27.762250882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.089449 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.089681 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.090021 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:22:37.590004269 +0000 UTC m=+27.763317621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.090302 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.090364 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.090606 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.090718 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.090909 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.092251 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.094913 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.097255 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.097594 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.098060 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.099206 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.100054 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.100205 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.100561 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.100818 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.101079 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.101690 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.103402 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.103457 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.103472 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.103537 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:37.603518251 +0000 UTC m=+27.776831813 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.103689 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.103833 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.104062 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.104267 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.105582 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.106203 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.107636 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.108253 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.108309 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.108352 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.108368 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.108419 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:37.608404121 +0000 UTC m=+27.781717673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.108520 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.108773 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.109085 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.109162 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.109578 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.109647 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.110076 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.110912 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.110969 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.111150 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.111166 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.111495 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.112498 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.112943 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.113277 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.115616 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.116470 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117007 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117475 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117731 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117739 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117806 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117816 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.117855 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.118291 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.119683 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.119763 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.119779 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.119835 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.120417 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.120487 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.120540 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.120804 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.120984 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.121446 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.129731 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.131358 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.139581 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.139995 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.141219 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.142459 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.145233 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.146191 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.148437 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.148796 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.149740 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.150147 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.151089 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.151881 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.153225 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.153958 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.155096 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.156723 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.156990 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.157735 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.159205 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.160830 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.162304 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.164128 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.164891 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.167695 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.168457 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.169788 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.170649 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.173082 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.173239 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.173848 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174482 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174598 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174705 4727 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174727 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174740 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174753 4727 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174764 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174775 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174785 4727 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174795 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174805 4727 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174817 4727 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174829 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174840 4727 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174851 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174861 4727 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174872 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174883 4727 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174893 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174902 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174912 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174922 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174931 4727 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174941 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174951 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174963 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174973 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174984 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.174995 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175007 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175019 4727 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175148 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175418 4727 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175611 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.175793 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176564 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176713 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176761 4727 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176842 4727 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176865 4727 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176881 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176896 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176910 4727 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176923 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176935 4727 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176947 4727 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176960 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176974 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176987 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.176999 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177011 4727 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177023 4727 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177034 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177045 4727 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177058 4727 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177069 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177081 4727 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177094 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177108 4727 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177120 4727 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177143 4727 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177156 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177169 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177181 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177193 4727 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177207 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177221 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177235 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177249 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177520 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.177790 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.178454 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.178688 4727 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.178882 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179138 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179170 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179186 4727 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179202 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179215 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179227 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179240 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179253 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179265 4727 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179278 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179294 4727 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179307 4727 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179319 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179345 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179358 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179406 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179419 4727 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179431 4727 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179443 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179457 4727 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179470 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179484 4727 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179500 4727 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179512 4727 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179524 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179536 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179548 4727 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179560 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179574 4727 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179585 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179598 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179611 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179623 4727 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179638 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179650 4727 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179664 4727 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179677 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179690 4727 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179700 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179711 4727 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179722 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179734 4727 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179746 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179757 4727 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179768 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179782 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179794 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179806 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179818 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179829 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179841 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179853 4727 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179866 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179877 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179889 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179901 4727 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179913 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179925 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179938 4727 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179951 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179964 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179977 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.179991 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180003 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180016 4727 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180030 4727 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180042 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180054 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180069 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180081 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180092 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180103 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180114 4727 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180126 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180137 4727 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180181 4727 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180195 4727 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180208 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180220 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180232 4727 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180243 4727 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180257 4727 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180269 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180282 4727 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180295 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180329 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180357 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.180575 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.181256 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.181516 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182379 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182406 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182416 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182428 4727 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182437 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182448 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.182671 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.183107 4727 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.183280 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.184197 4727 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.184432 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.184705 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.184878 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.185814 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.186918 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.187811 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.188886 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.189807 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.191108 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.192170 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.192795 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.193909 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.194555 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.195578 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.196129 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.196658 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.196702 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.211294 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.223182 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.312578 4727 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.321530 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.349702 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.361022 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.370258 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 10:22:37 crc kubenswrapper[4727]: W0929 10:22:37.391592 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3e3b4809370443ad3b14a0ec13c01472341ab42e86ff907f83282db716b02aad WatchSource:0}: Error finding container 3e3b4809370443ad3b14a0ec13c01472341ab42e86ff907f83282db716b02aad: Status 404 returned error can't find the container with id 3e3b4809370443ad3b14a0ec13c01472341ab42e86ff907f83282db716b02aad Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.588429 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.588540 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.588597 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:38.588583334 +0000 UTC m=+28.761896696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.688789 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.688923 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689055 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:22:38.689034092 +0000 UTC m=+28.862347454 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.689096 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689100 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689141 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689155 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689202 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689208 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:38.689189857 +0000 UTC m=+28.862503289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: I0929 10:22:37.689121 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689236 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:38.689228908 +0000 UTC m=+28.862542270 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689300 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689313 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689322 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:37 crc kubenswrapper[4727]: E0929 10:22:37.689362 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:38.689356481 +0000 UTC m=+28.862669843 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.311721 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mvx5d"] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.312009 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xwdcz"] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.312132 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w25jt"] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.312394 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.312457 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.312498 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.313397 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lqql2"] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.314023 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vg7wt"] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.314560 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.314669 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.315943 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.316307 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3e3b4809370443ad3b14a0ec13c01472341ab42e86ff907f83282db716b02aad"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.316599 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317023 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317067 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317211 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317324 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317356 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317427 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317454 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317701 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.317872 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.318039 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.318175 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.318445 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.319656 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.319711 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.319726 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"824b244273e7a877a8f9a501a2ff8ae17a4a9119938f3cad41dc10b25d4c6968"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.320517 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.320639 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.320952 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.321034 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.320676 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.322464 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.322935 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.323126 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"274dfa9e53949ee7f6d79a2f51fbc105966c4067d5693623c58359f34ffb70ed"} Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.324229 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.324558 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.332906 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.345281 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.360259 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.371048 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.381744 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.392783 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.395639 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-os-release\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.395669 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-conf-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.395697 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.395711 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.395734 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsb2b\" (UniqueName: \"kubernetes.io/projected/c74616e3-ccb3-453b-a7b7-b734d3203c06-kube-api-access-bsb2b\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396293 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd0ac62e-15bc-4508-a7da-414196360165-hosts-file\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396719 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396755 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396796 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-etc-kubernetes\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396880 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-netns\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396913 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-os-release\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396938 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-multus-certs\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396960 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.396986 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397032 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-binary-copy\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397141 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-system-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397183 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cni-binary-copy\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397223 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vlv\" (UniqueName: \"kubernetes.io/projected/cd0ac62e-15bc-4508-a7da-414196360165-kube-api-access-c6vlv\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397268 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397287 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397307 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397327 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-cnibin\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397366 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/791106af-2f47-4c80-9f20-e0bc4131f833-mcd-auth-proxy-config\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397467 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397598 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-socket-dir-parent\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397643 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-system-cni-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397665 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397690 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/791106af-2f47-4c80-9f20-e0bc4131f833-rootfs\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397733 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397754 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq6pj\" (UniqueName: \"kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397777 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-multus\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397799 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-daemon-config\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397820 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44csc\" (UniqueName: \"kubernetes.io/projected/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-kube-api-access-44csc\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397840 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397861 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397882 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397901 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cnibin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397917 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-hostroot\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.397976 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398046 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398102 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398136 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398169 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398221 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398250 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398282 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-k8s-cni-cncf-io\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398313 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-bin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398355 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-kubelet\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398386 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/791106af-2f47-4c80-9f20-e0bc4131f833-proxy-tls\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.398417 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj5ch\" (UniqueName: \"kubernetes.io/projected/791106af-2f47-4c80-9f20-e0bc4131f833-kube-api-access-fj5ch\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.411728 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.440103 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.460352 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.476325 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.490042 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499674 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-conf-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499720 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-os-release\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499744 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499778 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499812 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsb2b\" (UniqueName: \"kubernetes.io/projected/c74616e3-ccb3-453b-a7b7-b734d3203c06-kube-api-access-bsb2b\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499826 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-conf-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499839 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd0ac62e-15bc-4508-a7da-414196360165-hosts-file\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499914 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd0ac62e-15bc-4508-a7da-414196360165-hosts-file\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499926 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499950 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499968 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-etc-kubernetes\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499978 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.499985 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-netns\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500005 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-netns\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500023 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500031 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500049 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-os-release\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500040 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-os-release\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500075 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-multus-certs\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500097 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500122 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-binary-copy\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500135 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500154 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-system-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500177 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500187 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cni-binary-copy\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500216 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vlv\" (UniqueName: \"kubernetes.io/projected/cd0ac62e-15bc-4508-a7da-414196360165-kube-api-access-c6vlv\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500245 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500275 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-etc-kubernetes\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500272 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500306 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500312 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500381 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-cnibin\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500416 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/791106af-2f47-4c80-9f20-e0bc4131f833-mcd-auth-proxy-config\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500442 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500469 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/791106af-2f47-4c80-9f20-e0bc4131f833-rootfs\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500526 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500553 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-multus-certs\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500562 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-socket-dir-parent\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500596 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-system-cni-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500617 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-os-release\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500625 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-multus\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500655 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-cnibin\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500704 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-system-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500663 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-daemon-config\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500755 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44csc\" (UniqueName: \"kubernetes.io/projected/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-kube-api-access-44csc\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500802 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500838 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-binary-copy\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500839 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq6pj\" (UniqueName: \"kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500885 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500903 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cnibin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500918 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-hostroot\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500940 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500955 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500984 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501003 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501029 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501046 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501064 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501081 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-kubelet\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501096 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/791106af-2f47-4c80-9f20-e0bc4131f833-proxy-tls\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501115 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj5ch\" (UniqueName: \"kubernetes.io/projected/791106af-2f47-4c80-9f20-e0bc4131f833-kube-api-access-fj5ch\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501136 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501153 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501172 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-k8s-cni-cncf-io\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501186 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-bin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501254 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-bin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501382 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cni-binary-copy\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501604 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.500328 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501661 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-system-cni-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501717 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c74616e3-ccb3-453b-a7b7-b734d3203c06-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501871 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/791106af-2f47-4c80-9f20-e0bc4131f833-rootfs\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501949 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501963 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-cni-dir\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.501979 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/791106af-2f47-4c80-9f20-e0bc4131f833-mcd-auth-proxy-config\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502007 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-kubelet\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502031 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-socket-dir-parent\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502041 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-var-lib-cni-multus\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502040 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-cnibin\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502272 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502380 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-host-run-k8s-cni-cncf-io\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502401 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502312 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502308 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502356 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502379 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502274 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-hostroot\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502288 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502600 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.502958 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-multus-daemon-config\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.503015 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.503130 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.505796 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.506188 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c74616e3-ccb3-453b-a7b7-b734d3203c06-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.526234 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.540113 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.554811 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.566358 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.578656 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.591998 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.602899 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.603040 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.603169 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:40.603128408 +0000 UTC m=+30.776441770 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.615591 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.632263 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.644034 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.657154 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.667441 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.680763 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.697646 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.697758 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj5ch\" (UniqueName: \"kubernetes.io/projected/791106af-2f47-4c80-9f20-e0bc4131f833-kube-api-access-fj5ch\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.697803 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vlv\" (UniqueName: \"kubernetes.io/projected/cd0ac62e-15bc-4508-a7da-414196360165-kube-api-access-c6vlv\") pod \"node-resolver-mvx5d\" (UID: \"cd0ac62e-15bc-4508-a7da-414196360165\") " pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.698087 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/791106af-2f47-4c80-9f20-e0bc4131f833-proxy-tls\") pod \"machine-config-daemon-w25jt\" (UID: \"791106af-2f47-4c80-9f20-e0bc4131f833\") " pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.699541 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsb2b\" (UniqueName: \"kubernetes.io/projected/c74616e3-ccb3-453b-a7b7-b734d3203c06-kube-api-access-bsb2b\") pod \"multus-additional-cni-plugins-vg7wt\" (UID: \"c74616e3-ccb3-453b-a7b7-b734d3203c06\") " pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.699761 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq6pj\" (UniqueName: \"kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj\") pod \"ovnkube-node-lqql2\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.700375 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44csc\" (UniqueName: \"kubernetes.io/projected/9e3b98d9-a9fc-4bf3-8053-b7701c047d99-kube-api-access-44csc\") pod \"multus-xwdcz\" (UID: \"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\") " pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.703504 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.703615 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.703718 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:22:40.703682699 +0000 UTC m=+30.876996211 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.703746 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.703764 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.703778 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.703823 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:40.703806813 +0000 UTC m=+30.877120175 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.703840 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.703880 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704039 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704118 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:40.70409967 +0000 UTC m=+30.877413032 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704149 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704200 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704220 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:38 crc kubenswrapper[4727]: E0929 10:22:38.704307 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:40.704293806 +0000 UTC m=+30.877607358 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.932102 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.941029 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mvx5d" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.947486 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xwdcz" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.956506 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.961531 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:38 crc kubenswrapper[4727]: I0929 10:22:38.989768 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 29 10:22:39 crc kubenswrapper[4727]: W0929 10:22:39.003285 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc578ace_78c8_4d17_a556_c6d6ceb149a4.slice/crio-c69c1240554da2879557b426f455d8e7615dff9d27e752bdc81466afa2e60df8 WatchSource:0}: Error finding container c69c1240554da2879557b426f455d8e7615dff9d27e752bdc81466afa2e60df8: Status 404 returned error can't find the container with id c69c1240554da2879557b426f455d8e7615dff9d27e752bdc81466afa2e60df8 Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.015301 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.032913 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.051789 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.082692 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.112142 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:39 crc kubenswrapper[4727]: E0929 10:22:39.112354 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.112473 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:39 crc kubenswrapper[4727]: E0929 10:22:39.112538 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.113283 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:39 crc kubenswrapper[4727]: E0929 10:22:39.113378 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.118648 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.119942 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.143975 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.156881 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.171611 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.189915 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.204677 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.221275 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.245973 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.261862 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.279153 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.293782 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.315732 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.328803 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.328877 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"b74ce0ec7f7737fcaa6df2a8f1bb2ecdad71a1cfc14fb315d17958c6a3a53407"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.330277 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerStarted","Data":"d7583e0fae076d6f99bbba9682b43cc8442baacc94ce4d8fbd33e74a686372e6"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.331928 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerStarted","Data":"586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.331967 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerStarted","Data":"c3d35f1fbee272c6401997bbf8006957047de1fe42fbd4630a6bafa99590f879"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.333957 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mvx5d" event={"ID":"cd0ac62e-15bc-4508-a7da-414196360165","Type":"ContainerStarted","Data":"9a3f567e562591919dc18e4299f37e2417fee2f915ae13aee050f3a8f1926b89"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.334878 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.337829 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" exitCode=0 Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.338776 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.338817 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"c69c1240554da2879557b426f455d8e7615dff9d27e752bdc81466afa2e60df8"} Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.352091 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.371730 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.387505 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.418091 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.432930 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.449521 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.461904 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.480538 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.499638 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.514943 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.526171 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.543316 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.559791 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.576918 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.594106 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.611417 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.630156 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.644803 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.657397 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.673604 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.687806 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.700077 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.716763 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.738291 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.753174 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:39 crc kubenswrapper[4727]: I0929 10:22:39.767157 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:39Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.050665 4727 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.344833 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b" exitCode=0 Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.345177 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b"} Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.347288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4"} Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.348649 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mvx5d" event={"ID":"cd0ac62e-15bc-4508-a7da-414196360165","Type":"ContainerStarted","Data":"c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5"} Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.350253 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603"} Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.392090 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.417640 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.436106 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.457957 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.481043 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.496544 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.510373 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.527623 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.545990 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.559374 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.573915 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.595959 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.612637 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.626691 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.630788 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.630971 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.631097 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:44.631068497 +0000 UTC m=+34.804382039 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.642951 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.667265 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.687595 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.702747 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.716271 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.729766 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.732018 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732162 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:22:44.732142633 +0000 UTC m=+34.905455985 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.732232 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.732269 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.732297 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732415 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732442 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732448 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732459 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732493 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:44.732484222 +0000 UTC m=+34.905797584 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732493 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732517 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:44.732499712 +0000 UTC m=+34.905813074 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732525 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732539 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:40 crc kubenswrapper[4727]: E0929 10:22:40.732646 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:44.732581114 +0000 UTC m=+34.905894466 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.742606 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.765496 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.780180 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.795371 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.811353 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.837432 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.854411 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:40 crc kubenswrapper[4727]: I0929 10:22:40.877501 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:40Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.108246 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:41 crc kubenswrapper[4727]: E0929 10:22:41.108388 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.108448 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:41 crc kubenswrapper[4727]: E0929 10:22:41.108496 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.108553 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:41 crc kubenswrapper[4727]: E0929 10:22:41.141800 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.155406 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.174087 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.197027 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.214821 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.244966 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.291713 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.362851 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.364115 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943"} Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.364158 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f"} Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.367165 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerStarted","Data":"4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3"} Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.381703 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.394895 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.406961 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.421456 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.435036 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.449149 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.475162 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.492129 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.504403 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.521986 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.539476 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.555893 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.572688 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.595461 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.609036 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.624613 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.639702 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.653811 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.666630 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.704048 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:41 crc kubenswrapper[4727]: I0929 10:22:41.749885 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.077907 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-6lmbz"] Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.078552 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.081721 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.082135 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.082243 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.085300 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.096526 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.108590 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.119679 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.143869 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.152789 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slgjx\" (UniqueName: \"kubernetes.io/projected/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-kube-api-access-slgjx\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.152852 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-serviceca\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.152899 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-host\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.157829 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.171606 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.187151 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.207676 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.218700 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.233478 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.253591 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slgjx\" (UniqueName: \"kubernetes.io/projected/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-kube-api-access-slgjx\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.253630 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-serviceca\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.253657 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-host\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.253730 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-host\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.254638 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-serviceca\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.262646 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.289104 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slgjx\" (UniqueName: \"kubernetes.io/projected/02f5f1cc-410d-4f33-bf6c-8a4f06830e58-kube-api-access-slgjx\") pod \"node-ca-6lmbz\" (UID: \"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\") " pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.321397 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.362882 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.373644 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb"} Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.373691 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373"} Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.392123 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6lmbz" Sep 29 10:22:42 crc kubenswrapper[4727]: W0929 10:22:42.406490 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02f5f1cc_410d_4f33_bf6c_8a4f06830e58.slice/crio-e5d51f5a7537ed87a0852559a1919eeed2ab3311c4b99f51c6c99e0749f1af45 WatchSource:0}: Error finding container e5d51f5a7537ed87a0852559a1919eeed2ab3311c4b99f51c6c99e0749f1af45: Status 404 returned error can't find the container with id e5d51f5a7537ed87a0852559a1919eeed2ab3311c4b99f51c6c99e0749f1af45 Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.408421 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:42 crc kubenswrapper[4727]: I0929 10:22:42.444023 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.108170 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.108211 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.108254 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.108361 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.108478 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.108553 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.380678 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625"} Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.381868 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6lmbz" event={"ID":"02f5f1cc-410d-4f33-bf6c-8a4f06830e58","Type":"ContainerStarted","Data":"e5d51f5a7537ed87a0852559a1919eeed2ab3311c4b99f51c6c99e0749f1af45"} Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.660237 4727 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.662590 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.662646 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.662663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.662899 4727 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.670724 4727 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.671017 4727 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.672155 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.672200 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.672214 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.672235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.672247 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.690852 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.699867 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.699902 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.699913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.699931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.699943 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.712864 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.716682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.716708 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.716718 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.716733 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.716742 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.729333 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.733542 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.733596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.733608 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.733628 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.733642 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.746860 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.751819 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.751889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.751904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.751926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.751936 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.765519 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:43 crc kubenswrapper[4727]: E0929 10:22:43.765675 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.768499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.768534 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.768548 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.768573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.768588 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.870975 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.871015 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.871025 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.871042 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.871053 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.973354 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.973394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.973403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.973422 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:43 crc kubenswrapper[4727]: I0929 10:22:43.973434 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:43Z","lastTransitionTime":"2025-09-29T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.077402 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.077817 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.077827 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.077846 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.077869 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.179967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.180001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.180010 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.180025 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.180035 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.282230 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.282285 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.282300 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.282324 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.282358 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.383867 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.383908 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.383918 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.383936 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.383950 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.387929 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.389225 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6lmbz" event={"ID":"02f5f1cc-410d-4f33-bf6c-8a4f06830e58","Type":"ContainerStarted","Data":"9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.391119 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3" exitCode=0 Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.391154 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.413177 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.427985 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.444102 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.462783 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.481405 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.486801 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.486849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.486859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.486884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.486896 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.499070 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.517566 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.534418 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.547981 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.564196 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.579094 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.591971 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.592047 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.592062 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.592084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.592120 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.603132 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.635633 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.650824 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.665146 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.678229 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.678404 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.678455 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.678678 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:52.678634372 +0000 UTC m=+42.851947734 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.695134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.695166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.695176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.695191 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.695202 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.697291 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.710743 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.724867 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.738280 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.757156 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.768459 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.779573 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.779729 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779752 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:22:52.779731278 +0000 UTC m=+42.953044640 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.779809 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.779841 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779924 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779943 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779946 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779977 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.779997 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:52.779987295 +0000 UTC m=+42.953300657 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.780003 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.780033 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.780046 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.780013 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:52.780004715 +0000 UTC m=+42.953318077 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:44 crc kubenswrapper[4727]: E0929 10:22:44.780099 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:52.780083587 +0000 UTC m=+42.953396949 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.784081 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797598 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797856 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797868 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.797878 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.812529 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.825626 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.841418 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.856867 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.872040 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.883363 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.900392 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.900446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.900458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.900476 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:44 crc kubenswrapper[4727]: I0929 10:22:44.900489 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:44Z","lastTransitionTime":"2025-09-29T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.002555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.002596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.002605 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.002620 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.002630 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.105209 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.105269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.105284 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.105310 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.105324 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.107669 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.107691 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:45 crc kubenswrapper[4727]: E0929 10:22:45.107831 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.107981 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:45 crc kubenswrapper[4727]: E0929 10:22:45.108123 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:45 crc kubenswrapper[4727]: E0929 10:22:45.108418 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.208275 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.208386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.208411 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.208447 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.208470 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.310943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.310991 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.311006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.311023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.311034 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.413222 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.413252 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.413260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.413272 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.413281 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.515290 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.515326 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.515360 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.515377 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.515387 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.617894 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.617931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.617939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.617954 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.617964 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.720099 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.720164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.720186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.720216 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.720238 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.822937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.822978 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.822990 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.823009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.823024 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.926684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.926740 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.926752 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.926770 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:45 crc kubenswrapper[4727]: I0929 10:22:45.926783 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:45Z","lastTransitionTime":"2025-09-29T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.028916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.028962 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.028974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.028993 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.029005 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.131081 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.131177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.131194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.131252 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.131272 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.233831 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.233903 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.233919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.233941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.233957 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.337404 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.337482 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.337494 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.337511 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.337523 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.403265 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22" exitCode=0 Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.403304 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.416223 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.431761 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.440086 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.440154 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.440167 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.440194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.440206 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.447207 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.463589 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.481324 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.495776 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.515374 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.529212 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542123 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542564 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542591 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542600 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.542626 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.555931 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.575231 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.579883 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.592900 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.612725 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.629842 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.645289 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.645674 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.645693 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.645715 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.645731 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.646226 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.659860 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.707661 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.729279 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.743651 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.747773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.747836 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.747851 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.747868 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.747901 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.757224 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.772291 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.790103 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.801417 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.814045 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.827058 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.843710 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.851006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.851040 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.851051 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.851069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.851081 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.876559 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.901290 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.919972 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.935829 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.953423 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.953463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.953472 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.953497 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:46 crc kubenswrapper[4727]: I0929 10:22:46.953506 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:46Z","lastTransitionTime":"2025-09-29T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.055782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.055832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.055840 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.055856 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.055866 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.107606 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.107688 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:47 crc kubenswrapper[4727]: E0929 10:22:47.107722 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:47 crc kubenswrapper[4727]: E0929 10:22:47.107823 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.107912 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:47 crc kubenswrapper[4727]: E0929 10:22:47.108023 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.158737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.158797 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.158833 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.158860 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.158876 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.261206 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.261240 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.261250 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.261264 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.261276 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.363712 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.363758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.363768 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.363785 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.363800 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.409583 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.412248 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b" exitCode=0 Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.412281 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.430995 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.446438 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.463943 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.467910 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.468134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.468151 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.468171 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.468183 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.477148 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.491713 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.506520 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.520509 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.534878 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.555056 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.567844 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.570900 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.570931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.570941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.570957 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.570971 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.580574 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.594090 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.611930 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.623492 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.636154 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:47Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.673970 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.674008 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.674016 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.674033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.674044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.776719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.776759 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.776770 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.776785 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.776796 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.878871 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.878925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.878947 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.878967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.878982 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.982427 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.982464 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.982473 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.982491 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:47 crc kubenswrapper[4727]: I0929 10:22:47.982501 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:47Z","lastTransitionTime":"2025-09-29T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.084306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.084374 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.084386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.084402 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.084412 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.187372 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.187418 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.187428 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.187444 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.187469 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.293783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.293818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.293828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.293843 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.293853 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.396252 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.396304 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.396315 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.396348 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.396361 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.429497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerStarted","Data":"b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.441276 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.454186 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.470786 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.484264 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.498324 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.500700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.500746 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.500762 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.500780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.500793 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.519373 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.535450 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.547633 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.568360 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.581088 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.593736 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.603087 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.603112 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.603121 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.603134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.603143 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.607798 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.630305 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.646996 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.662825 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:48Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.710689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.710728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.710736 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.710751 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.710761 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.814412 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.814446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.814456 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.814469 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.814479 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.917309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.917381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.917415 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.917440 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:48 crc kubenswrapper[4727]: I0929 10:22:48.917454 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:48Z","lastTransitionTime":"2025-09-29T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.019665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.019750 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.019764 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.019783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.019795 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.108654 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.108771 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:49 crc kubenswrapper[4727]: E0929 10:22:49.108846 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.108892 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:49 crc kubenswrapper[4727]: E0929 10:22:49.108964 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:49 crc kubenswrapper[4727]: E0929 10:22:49.109125 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.122360 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.122538 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.122596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.122695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.122903 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.226041 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.226124 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.226149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.226176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.226194 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.328757 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.328834 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.328859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.328890 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.328911 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.431619 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.431681 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.431700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.431723 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.431739 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.534243 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.534290 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.534301 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.534320 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.534358 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.638408 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.638459 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.638470 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.638488 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.638499 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.742048 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.742455 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.742469 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.742486 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.742500 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.844294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.844350 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.844363 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.844379 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.844390 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.946735 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.946771 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.946779 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.946792 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:49 crc kubenswrapper[4727]: I0929 10:22:49.946803 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:49Z","lastTransitionTime":"2025-09-29T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.049493 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.049570 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.049587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.049614 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.049636 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.152540 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.152579 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.152590 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.152606 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.152618 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.255066 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.255116 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.255127 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.255148 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.255163 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.357866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.357914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.357926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.357945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.357956 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.442691 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.443085 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.461869 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.461925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.461942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.461974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.461990 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.462190 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.476808 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.480625 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.501687 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.519274 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.537616 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.552353 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.564866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.564974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.564990 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.565014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.565031 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.570149 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.589647 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.607139 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.624241 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.639634 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.664107 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.668101 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.668162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.668175 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.668200 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.668213 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.680261 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.694878 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.709920 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.739837 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.755165 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.769917 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.770989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.771044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.771059 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.771084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.771103 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.798485 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.813234 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.826112 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.839752 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.857691 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.869666 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.873644 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.873739 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.873815 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.873907 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.873980 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.882392 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.903889 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.914900 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.924741 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.933824 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.947273 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:50Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.979646 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.979873 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.980002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.980089 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:50 crc kubenswrapper[4727]: I0929 10:22:50.980233 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:50Z","lastTransitionTime":"2025-09-29T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.082452 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.082481 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.082489 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.082501 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.082529 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.098187 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr"] Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.098726 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.102911 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.102995 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.107628 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.107659 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:51 crc kubenswrapper[4727]: E0929 10:22:51.107739 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.107638 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:51 crc kubenswrapper[4727]: E0929 10:22:51.107845 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:51 crc kubenswrapper[4727]: E0929 10:22:51.107903 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.115127 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.125438 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.138501 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.157700 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.171035 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.184511 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.184561 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.184574 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.184594 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.184608 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.185827 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.201903 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.215226 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.231864 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.247462 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.253113 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfmmt\" (UniqueName: \"kubernetes.io/projected/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-kube-api-access-rfmmt\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.253173 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.253194 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-env-overrides\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.253208 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.259586 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.269926 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.286789 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.286824 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.286834 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.286851 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.286863 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.287270 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.298763 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.312789 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.323471 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.332848 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.346077 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.354379 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-env-overrides\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.354414 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.354451 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfmmt\" (UniqueName: \"kubernetes.io/projected/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-kube-api-access-rfmmt\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.354486 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.355569 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-env-overrides\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.355606 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.356507 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.359864 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.370698 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfmmt\" (UniqueName: \"kubernetes.io/projected/9cdf5dfa-ea0f-48e8-9f56-df22343f7d58-kube-api-access-rfmmt\") pod \"ovnkube-control-plane-749d76644c-d8zdr\" (UID: \"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.371058 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.388614 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.388857 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.388915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.388973 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.389027 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.393895 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.407150 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.413062 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.417981 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.442542 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.448519 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f" exitCode=0 Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.448583 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.449766 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" event={"ID":"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58","Type":"ContainerStarted","Data":"3c3324952ac16eee6cb8446acff5aeb153e159e4450ecd355c6994c5587eb545"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.450381 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.450743 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.471253 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.476015 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.482926 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.491658 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.491700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.491728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.491745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.491757 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.498854 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.514097 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.524942 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.533659 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.546583 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.557969 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.569389 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.579713 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.591269 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.593841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.593879 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.593887 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.593901 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.593911 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.603171 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.625696 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.640382 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.653910 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.667863 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.683467 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.696293 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.696324 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.696348 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.696366 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.696375 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.697626 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.710485 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.723829 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.741620 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.754850 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.765471 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.782149 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.799018 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.799057 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.799067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.799083 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.799094 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.901163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.901200 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.901208 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.901221 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:51 crc kubenswrapper[4727]: I0929 10:22:51.901230 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:51Z","lastTransitionTime":"2025-09-29T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.004469 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.004510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.004529 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.004546 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.004556 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.106800 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.106855 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.106868 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.106887 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.106898 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.210976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.211048 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.211073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.211103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.211129 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.314679 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.314740 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.314758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.314781 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.314797 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.416916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.416974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.416992 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.417020 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.417040 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.454498 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" event={"ID":"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58","Type":"ContainerStarted","Data":"136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.519729 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.520056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.520070 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.520089 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.520105 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.582860 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gttbj"] Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.583353 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.583420 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.608150 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623300 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623351 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623361 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623376 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623384 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.623475 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.636758 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.667322 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.698173 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.713853 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726011 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726100 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726147 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726192 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.726204 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.741520 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.757421 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.769506 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btrmc\" (UniqueName: \"kubernetes.io/projected/59055eb3-6cad-4335-b100-39955f1c0500-kube-api-access-btrmc\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.769606 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.769646 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.769912 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.770121 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.77003929 +0000 UTC m=+58.943352652 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.775372 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.786237 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.803038 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.816204 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.826930 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.828569 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.828612 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.828624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.828640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.828651 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.839851 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.851187 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.866080 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:52Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.870551 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.870685 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.870721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.870758 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btrmc\" (UniqueName: \"kubernetes.io/projected/59055eb3-6cad-4335-b100-39955f1c0500-kube-api-access-btrmc\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870787 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870851 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.870818708 +0000 UTC m=+59.044132080 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870895 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.870884209 +0000 UTC m=+59.044197661 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870869 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870934 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870954 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870959 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.870998 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.870990012 +0000 UTC m=+59.044303454 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.870792 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.871043 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:53.371020503 +0000 UTC m=+43.544333885 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.871078 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.871238 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.871260 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.871277 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:52 crc kubenswrapper[4727]: E0929 10:22:52.871316 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.871302581 +0000 UTC m=+59.044615953 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.889979 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btrmc\" (UniqueName: \"kubernetes.io/projected/59055eb3-6cad-4335-b100-39955f1c0500-kube-api-access-btrmc\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.930972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.931006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.931019 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.931033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:52 crc kubenswrapper[4727]: I0929 10:22:52.931044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:52Z","lastTransitionTime":"2025-09-29T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.034223 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.034260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.034268 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.034282 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.034291 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.108489 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.108551 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.108517 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.108658 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.108833 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.108999 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.136745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.136807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.136825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.136845 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.136860 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.240164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.240211 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.240227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.240250 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.240267 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.343845 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.343904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.343922 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.343945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.343962 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.376983 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.377162 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.377230 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:54.377214611 +0000 UTC m=+44.550527973 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.447616 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.447659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.447720 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.447735 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.447744 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.465246 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerStarted","Data":"caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.549962 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.550007 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.550019 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.550038 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.550054 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.653397 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.653450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.653460 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.653477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.653488 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.756912 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.756966 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.756974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.756992 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.757002 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.860176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.860241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.860256 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.860275 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.860287 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.907991 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.908091 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.908118 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.908154 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.908179 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.931279 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:53Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.938677 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.938751 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.938774 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.938807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.938826 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.962506 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:53Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.969486 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.969533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.969550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.969568 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.969581 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:53 crc kubenswrapper[4727]: E0929 10:22:53.991054 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:53Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.997096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.997175 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.997193 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.997225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:53 crc kubenswrapper[4727]: I0929 10:22:53.997251 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:53Z","lastTransitionTime":"2025-09-29T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.026529 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.033015 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.033096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.033117 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.033146 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.033167 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.054427 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.054637 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.057598 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.057683 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.057707 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.057745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.057764 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.107736 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.108081 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.162205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.162281 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.162302 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.162333 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.162392 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.266562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.266629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.266647 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.266678 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.266697 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.370857 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.370927 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.370942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.370966 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.370982 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.386749 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.386958 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:54 crc kubenswrapper[4727]: E0929 10:22:54.387044 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:22:56.387019409 +0000 UTC m=+46.560332781 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.473806 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.473929 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.473958 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.474036 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.474061 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.474900 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" event={"ID":"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58","Type":"ContainerStarted","Data":"afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.481173 4727 generic.go:334] "Generic (PLEG): container finished" podID="c74616e3-ccb3-453b-a7b7-b734d3203c06" containerID="caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e" exitCode=0 Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.481247 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerDied","Data":"caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.511273 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.528489 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.542988 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.564683 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.580621 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.580672 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.580686 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.580710 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.580725 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.582769 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.598220 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.613544 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.634624 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.646838 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.658931 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.671925 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683182 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683892 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683957 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.683989 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.697478 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.711220 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.722412 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.737471 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.750697 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.766003 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.778257 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.786239 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.786273 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.786283 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.786300 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.786310 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.788985 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.804690 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.823738 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.834690 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.845771 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.857944 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.868024 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.882314 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.888739 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.888768 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.888779 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.888794 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.888807 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.901090 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.914411 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.929272 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.943317 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.962856 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.975681 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.986872 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:54Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.991172 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.991412 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.991500 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.991572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:54 crc kubenswrapper[4727]: I0929 10:22:54.991645 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:54Z","lastTransitionTime":"2025-09-29T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.093554 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.093589 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.093602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.093619 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.093631 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.108384 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.108490 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.108412 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:55 crc kubenswrapper[4727]: E0929 10:22:55.108592 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:55 crc kubenswrapper[4727]: E0929 10:22:55.108706 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:55 crc kubenswrapper[4727]: E0929 10:22:55.108787 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.196017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.196064 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.196073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.196103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.196113 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.298052 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.298101 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.298113 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.298129 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.298138 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.400695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.400779 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.400807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.400840 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.400868 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.487682 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" event={"ID":"c74616e3-ccb3-453b-a7b7-b734d3203c06","Type":"ContainerStarted","Data":"ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503180 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503815 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503829 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.503841 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.515370 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.528808 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.543942 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.566155 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.581028 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.594858 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.606495 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.606548 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.606566 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.606592 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.606609 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.610676 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.625437 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.644254 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.658107 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.694532 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.707970 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.710131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.710161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.710170 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.710186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.710196 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.724002 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.740962 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.755765 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.770197 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:55Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.812787 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.812830 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.812838 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.812854 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.812864 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.916094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.916152 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.916161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.916179 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:55 crc kubenswrapper[4727]: I0929 10:22:55.916189 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:55Z","lastTransitionTime":"2025-09-29T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.020709 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.020762 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.020777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.020794 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.020806 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.107916 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:56 crc kubenswrapper[4727]: E0929 10:22:56.108120 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.123986 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.124073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.124101 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.124137 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.124158 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.227808 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.227847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.227858 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.227875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.227885 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.330354 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.330399 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.330410 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.330426 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.330441 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.408146 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:56 crc kubenswrapper[4727]: E0929 10:22:56.408303 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:56 crc kubenswrapper[4727]: E0929 10:22:56.408392 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:00.40837322 +0000 UTC m=+50.581686582 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.432850 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.432888 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.432901 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.432917 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.432928 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.534817 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.535172 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.535184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.535203 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.535214 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.638814 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.638888 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.638905 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.638931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.638948 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.741513 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.741554 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.741566 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.741584 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.741596 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.844582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.844613 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.844624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.844641 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.844652 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.947830 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.947857 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.947866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.947880 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:56 crc kubenswrapper[4727]: I0929 10:22:56.947890 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:56Z","lastTransitionTime":"2025-09-29T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.051128 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.051177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.051191 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.051214 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.051225 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.108236 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.108314 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:57 crc kubenswrapper[4727]: E0929 10:22:57.108394 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:57 crc kubenswrapper[4727]: E0929 10:22:57.108552 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.108607 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:57 crc kubenswrapper[4727]: E0929 10:22:57.108690 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.154936 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.154999 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.155015 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.155038 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.155053 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.258460 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.258526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.258537 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.258558 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.258572 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.361722 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.361797 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.361810 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.361875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.361891 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.465125 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.465173 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.465182 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.465199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.465211 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.496405 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/0.log" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.499800 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f" exitCode=1 Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.499835 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.500501 4727 scope.go:117] "RemoveContainer" containerID="31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.521787 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.534136 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.552323 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569289 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569356 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569373 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569392 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569403 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.569952 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.584037 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.600445 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.614837 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.634652 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.654239 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.668695 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.671302 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.671360 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.671370 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.671386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.671398 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.689719 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.709739 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.730639 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.748118 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.770194 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:56Z\\\",\\\"message\\\":\\\"IP event handler 8 for removal\\\\nI0929 10:22:56.858220 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 10:22:56.858231 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 10:22:56.858258 6010 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0929 10:22:56.858388 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0929 10:22:56.858402 6010 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 10:22:56.858445 6010 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.858783 6010 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858952 6010 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858992 6010 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.859256 6010 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.774949 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.774980 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.774992 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.775006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.775016 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.785266 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.800384 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:57Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.878119 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.878403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.878486 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.878576 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.878849 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.981177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.981215 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.981224 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.981239 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:57 crc kubenswrapper[4727]: I0929 10:22:57.981249 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:57Z","lastTransitionTime":"2025-09-29T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.085910 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.086295 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.086306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.086323 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.086347 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.108253 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:22:58 crc kubenswrapper[4727]: E0929 10:22:58.108458 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.188636 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.188732 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.188764 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.188797 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.188820 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.292904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.292958 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.292971 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.292995 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.293009 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.395731 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.395771 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.395780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.395796 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.395806 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.503352 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.503395 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.503407 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.503431 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.503444 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.506781 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/0.log" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.510569 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.510962 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.524950 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.538871 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.551685 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.564472 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.586240 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.600711 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.605458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.605499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.605510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.605526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.605537 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.613846 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.625143 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.638508 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.650223 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.669792 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.683779 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.700419 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.707767 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.707802 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.707812 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.707825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.707835 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.721266 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:56Z\\\",\\\"message\\\":\\\"IP event handler 8 for removal\\\\nI0929 10:22:56.858220 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 10:22:56.858231 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 10:22:56.858258 6010 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0929 10:22:56.858388 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0929 10:22:56.858402 6010 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 10:22:56.858445 6010 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.858783 6010 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858952 6010 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858992 6010 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.859256 6010 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.733092 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.745459 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.763101 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:58Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.809995 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.810044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.810056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.810067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.810076 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.912877 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.913241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.913471 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.914134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:58 crc kubenswrapper[4727]: I0929 10:22:58.914292 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:58Z","lastTransitionTime":"2025-09-29T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.016763 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.016802 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.016813 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.016828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.016838 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.108276 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:22:59 crc kubenswrapper[4727]: E0929 10:22:59.108482 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.108323 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:22:59 crc kubenswrapper[4727]: E0929 10:22:59.108574 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.108292 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:22:59 crc kubenswrapper[4727]: E0929 10:22:59.108674 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.119588 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.119784 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.119844 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.119903 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.120017 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.222864 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.222904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.222915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.222933 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.222946 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.325876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.325942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.325965 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.325998 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.326022 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.428528 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.428579 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.428588 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.428602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.428613 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.515596 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/1.log" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.517078 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/0.log" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.520530 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260" exitCode=1 Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.520591 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.520678 4727 scope.go:117] "RemoveContainer" containerID="31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.522206 4727 scope.go:117] "RemoveContainer" containerID="ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260" Sep 29 10:22:59 crc kubenswrapper[4727]: E0929 10:22:59.522552 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.531868 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.532091 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.532190 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.532279 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.532394 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.541077 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.555955 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.568880 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.579572 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.592680 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.607418 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.622154 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.634884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.634915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.634924 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.634938 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.634948 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.645361 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.660328 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.674204 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.689755 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.714068 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31951d554f44a3081bfb80d341bfe52f367456910d428cc27576e4600bd02f6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:56Z\\\",\\\"message\\\":\\\"IP event handler 8 for removal\\\\nI0929 10:22:56.858220 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0929 10:22:56.858231 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0929 10:22:56.858258 6010 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0929 10:22:56.858388 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0929 10:22:56.858402 6010 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0929 10:22:56.858445 6010 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.858783 6010 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858952 6010 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 10:22:56.858992 6010 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 10:22:56.859256 6010 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.726983 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.738448 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.738502 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.738530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.738552 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.738567 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.744657 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.767198 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.782368 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.801237 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:22:59Z is after 2025-08-24T17:21:41Z" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.840764 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.840809 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.840819 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.840834 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.840845 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.943636 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.943684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.943695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.943711 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:22:59 crc kubenswrapper[4727]: I0929 10:22:59.943723 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:22:59Z","lastTransitionTime":"2025-09-29T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.046347 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.046386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.046394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.046410 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.046421 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.108029 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:00 crc kubenswrapper[4727]: E0929 10:23:00.108183 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.149178 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.149221 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.149230 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.149246 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.149259 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.252585 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.252627 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.252640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.252659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.252675 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.356088 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.356142 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.356161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.356186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.356202 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.445379 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:00 crc kubenswrapper[4727]: E0929 10:23:00.445547 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:00 crc kubenswrapper[4727]: E0929 10:23:00.445613 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:08.445594099 +0000 UTC m=+58.618907451 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.459131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.459164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.459173 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.459186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.459198 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.525518 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/1.log" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.528539 4727 scope.go:117] "RemoveContainer" containerID="ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260" Sep 29 10:23:00 crc kubenswrapper[4727]: E0929 10:23:00.528839 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.540952 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.553227 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.561530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.561593 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.561614 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.561642 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.561661 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.567210 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.592839 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.629237 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.643655 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.654522 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.664073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.664120 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.664132 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.664149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.664162 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.668136 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.680627 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.689128 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.692122 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.696071 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.704717 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.722210 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.731208 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.741258 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.752525 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.761545 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.765874 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.765925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.765937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.765955 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.765968 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.776263 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.787129 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.800315 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.810088 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.826273 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.839828 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.852476 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.867062 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.872667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.872718 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.872731 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.872754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.872767 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.884855 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.914055 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.927358 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.937990 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.952507 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.965424 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.975482 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.975512 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.975520 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.975532 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.975541 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:00Z","lastTransitionTime":"2025-09-29T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.977968 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:00 crc kubenswrapper[4727]: I0929 10:23:00.994080 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:00Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.014097 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.026058 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.038844 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.078238 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.078596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.078692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.078781 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.078881 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.107586 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.107653 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:01 crc kubenswrapper[4727]: E0929 10:23:01.107721 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.107809 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:01 crc kubenswrapper[4727]: E0929 10:23:01.107918 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:01 crc kubenswrapper[4727]: E0929 10:23:01.108005 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.125384 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.142050 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.158506 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.173411 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.181428 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.181474 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.181484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.181501 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.181511 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.195616 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.208239 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.219149 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.232730 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.246030 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.259104 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.272982 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.284394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.284462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.284477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.284498 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.284513 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.291226 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.307439 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.319435 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.335706 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.348554 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.360688 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.374517 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:01Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.387484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.387550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.387562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.387579 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.387589 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.489952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.490002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.490014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.490030 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.490044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.592241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.592294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.592306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.592321 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.592329 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.694095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.694137 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.694174 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.694189 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.694199 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.796039 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.796671 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.796750 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.796827 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.796891 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.899356 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.899613 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.899725 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.899809 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:01 crc kubenswrapper[4727]: I0929 10:23:01.899886 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:01Z","lastTransitionTime":"2025-09-29T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.002725 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.002758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.002767 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.002782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.002793 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.104977 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.105220 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.105282 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.105376 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.105448 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.108211 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:02 crc kubenswrapper[4727]: E0929 10:23:02.108325 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.207891 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.207934 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.207945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.207959 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.207970 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.311012 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.311059 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.311070 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.311087 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.311098 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.414172 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.414216 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.414227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.414246 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.414257 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.516937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.516986 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.517001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.517017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.517028 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.619749 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.619796 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.619824 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.619849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.619864 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.723388 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.723438 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.723458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.723481 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.723497 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.826169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.826241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.826276 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.826309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.826330 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.929718 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.929770 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.929782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.929804 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:02 crc kubenswrapper[4727]: I0929 10:23:02.929825 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:02Z","lastTransitionTime":"2025-09-29T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.033937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.033973 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.033985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.034027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.034038 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.108557 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.108558 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:03 crc kubenswrapper[4727]: E0929 10:23:03.108676 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:03 crc kubenswrapper[4727]: E0929 10:23:03.108850 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.108951 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:03 crc kubenswrapper[4727]: E0929 10:23:03.109081 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.136932 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.137198 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.137265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.137349 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.137423 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.241138 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.241516 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.241986 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.242134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.242279 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.345614 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.345893 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.345997 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.346097 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.346183 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.449037 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.449572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.449669 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.449773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.449864 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.563403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.563489 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.563504 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.563533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.563553 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.667377 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.667417 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.667428 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.667444 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.667452 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.770618 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.770719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.770739 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.770837 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.770929 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.873598 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.873645 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.873660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.873681 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.873696 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.976117 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.976172 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.976189 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.976209 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:03 crc kubenswrapper[4727]: I0929 10:23:03.976223 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:03Z","lastTransitionTime":"2025-09-29T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.079386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.079458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.079469 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.079489 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.079500 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.107935 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.108086 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.181651 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.181691 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.181701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.181717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.181730 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.283948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.283986 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.284001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.284017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.284029 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.354158 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.354211 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.354221 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.354235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.354245 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.367704 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:04Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.371771 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.371822 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.371849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.371872 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.371887 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.383221 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:04Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.387036 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.387073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.387082 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.387098 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.387108 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.398206 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:04Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.402290 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.402352 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.402364 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.402382 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.402393 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.418164 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:04Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.424850 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.424888 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.424899 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.424913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.424922 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.437046 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:04Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:04 crc kubenswrapper[4727]: E0929 10:23:04.437164 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.438537 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.438560 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.438568 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.438580 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.438588 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.540527 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.540557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.540565 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.540579 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.540589 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.642521 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.642561 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.642571 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.642588 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.642599 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.744181 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.744222 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.744232 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.744248 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.744261 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.847350 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.847385 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.847395 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.847410 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.847425 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.950067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.950151 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.950163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.950180 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:04 crc kubenswrapper[4727]: I0929 10:23:04.950190 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:04Z","lastTransitionTime":"2025-09-29T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.053423 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.053488 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.053501 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.053522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.053535 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.107724 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.107799 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:05 crc kubenswrapper[4727]: E0929 10:23:05.107855 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.107811 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:05 crc kubenswrapper[4727]: E0929 10:23:05.107964 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:05 crc kubenswrapper[4727]: E0929 10:23:05.108004 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.156688 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.156761 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.156791 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.156816 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.156830 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.259633 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.259682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.259692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.259705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.259714 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.362129 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.362176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.362186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.362202 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.362213 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.464334 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.464410 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.464420 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.464435 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.464444 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.566582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.566657 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.566667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.566682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.566691 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.669506 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.669547 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.669557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.669572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.669581 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.772728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.772780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.772797 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.772820 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.772835 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.875295 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.875352 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.875362 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.875380 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.875390 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.977475 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.977519 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.977531 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.977550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:05 crc kubenswrapper[4727]: I0929 10:23:05.977561 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:05Z","lastTransitionTime":"2025-09-29T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.079700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.079758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.079772 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.079790 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.079803 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.108308 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:06 crc kubenswrapper[4727]: E0929 10:23:06.108501 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.182297 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.182392 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.182410 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.182427 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.182439 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.284913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.284946 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.284957 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.284983 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.284996 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.387501 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.387542 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.387558 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.387581 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.387592 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.489909 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.489948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.489958 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.489974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.489984 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.592168 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.592227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.592236 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.592253 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.592262 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.695022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.695084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.695099 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.695116 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.695128 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.797766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.797811 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.797821 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.797837 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.797847 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.901483 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.901532 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.901545 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.901566 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:06 crc kubenswrapper[4727]: I0929 10:23:06.901582 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:06Z","lastTransitionTime":"2025-09-29T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.004129 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.004178 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.004190 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.004206 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.004218 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.106826 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.106893 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.106906 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.106926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.106939 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.107728 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.107755 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.107872 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:07 crc kubenswrapper[4727]: E0929 10:23:07.107863 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:07 crc kubenswrapper[4727]: E0929 10:23:07.107946 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:07 crc kubenswrapper[4727]: E0929 10:23:07.108071 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.209616 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.209652 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.209665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.209682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.209693 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.312171 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.312242 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.312254 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.312271 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.312283 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.415593 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.415624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.415632 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.415647 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.415657 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.518022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.518119 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.518138 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.518162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.518179 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.621230 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.621297 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.621324 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.621361 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.621371 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.724448 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.724517 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.724529 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.724548 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.724560 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.826838 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.826885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.826895 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.826913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.826924 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.929774 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.929852 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.929865 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.929880 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:07 crc kubenswrapper[4727]: I0929 10:23:07.929893 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:07Z","lastTransitionTime":"2025-09-29T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.032996 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.033043 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.033054 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.033072 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.033084 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.108232 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.108469 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.135102 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.135161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.135171 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.135187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.135198 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.238033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.238094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.238111 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.238137 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.238156 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.341504 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.341550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.341561 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.341576 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.341587 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.445136 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.445187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.445198 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.445216 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.445226 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.536930 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.537164 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.537370 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:24.537352017 +0000 UTC m=+74.710665379 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.548009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.548044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.548053 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.548067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.548077 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.650097 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.650141 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.650150 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.650164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.650172 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.752956 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.752998 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.753009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.753027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.753038 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.841000 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.841191 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.841331 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:40.841300783 +0000 UTC m=+91.014614185 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.855964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.856020 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.856033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.856051 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.856063 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.942491 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.942673 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942734 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:23:40.942703987 +0000 UTC m=+91.116017349 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.942795 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942845 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.942853 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942870 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942889 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942952 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942974 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942985 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.943009 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.942956 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:40.942933353 +0000 UTC m=+91.116246755 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.943034 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:40.943026765 +0000 UTC m=+91.116340117 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:08 crc kubenswrapper[4727]: E0929 10:23:08.943046 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:40.943041046 +0000 UTC m=+91.116354408 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.958377 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.958428 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.958447 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.958464 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:08 crc kubenswrapper[4727]: I0929 10:23:08.958474 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:08Z","lastTransitionTime":"2025-09-29T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.061121 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.061442 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.061451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.061468 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.061477 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.107790 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.107893 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.108017 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:09 crc kubenswrapper[4727]: E0929 10:23:09.108015 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:09 crc kubenswrapper[4727]: E0929 10:23:09.108188 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:09 crc kubenswrapper[4727]: E0929 10:23:09.108305 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.163642 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.163692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.163702 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.163718 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.163728 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.265780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.265829 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.265841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.265855 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.265863 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.369730 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.369814 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.369833 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.369885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.369901 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.472916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.472965 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.472983 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.473005 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.473020 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.575812 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.575884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.575900 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.575924 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.575943 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.678520 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.678562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.678571 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.678588 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.678598 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.780894 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.780937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.780952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.780973 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.780986 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.883642 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.883677 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.883688 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.883705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.883714 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.986889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.986934 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.987010 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.987028 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:09 crc kubenswrapper[4727]: I0929 10:23:09.987037 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:09Z","lastTransitionTime":"2025-09-29T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.090573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.090673 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.090717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.090753 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.090779 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.108070 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:10 crc kubenswrapper[4727]: E0929 10:23:10.108399 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.194746 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.194818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.194839 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.194869 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.194890 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.297430 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.297475 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.297486 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.297502 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.297513 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.400012 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.400056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.400073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.400090 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.400101 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.501837 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.501888 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.501899 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.501919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.501931 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.605308 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.605382 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.605395 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.605411 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.605422 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.707849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.707909 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.707919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.707939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.707951 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.810235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.810284 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.810330 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.810371 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.810382 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.912948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.913021 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.913034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.913050 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:10 crc kubenswrapper[4727]: I0929 10:23:10.913062 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:10Z","lastTransitionTime":"2025-09-29T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.015363 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.015407 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.015418 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.015432 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.015442 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.107744 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.107791 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.107812 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:11 crc kubenswrapper[4727]: E0929 10:23:11.108010 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:11 crc kubenswrapper[4727]: E0929 10:23:11.108187 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:11 crc kubenswrapper[4727]: E0929 10:23:11.108834 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.117783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.118159 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.118859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.118932 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.118959 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.133139 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.148599 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.162627 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.186635 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.196794 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.211917 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.222191 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.222235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.222249 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.222266 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.222280 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.227404 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.240263 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.254708 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.266919 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.282262 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.293880 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.306604 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.317370 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.324987 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.325023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.325033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.325048 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.325059 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.331194 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.343162 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.356532 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.371625 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:11Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.427046 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.427094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.427105 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.427120 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.427130 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.529964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.530004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.530016 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.530031 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.530042 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.632844 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.632880 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.632889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.632903 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.632913 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.737075 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.737118 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.737131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.737147 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.737158 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.839004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.839035 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.839045 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.839059 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.839069 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.941593 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.941632 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.941641 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.941654 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:11 crc kubenswrapper[4727]: I0929 10:23:11.941664 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:11Z","lastTransitionTime":"2025-09-29T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.043604 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.043659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.043674 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.043706 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.043722 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.107383 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:12 crc kubenswrapper[4727]: E0929 10:23:12.107674 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.108494 4727 scope.go:117] "RemoveContainer" containerID="ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.146912 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.147326 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.147572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.147734 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.147868 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.250497 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.250533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.250543 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.250557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.250566 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.353755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.353806 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.353818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.353838 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.353852 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.456655 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.456731 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.456752 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.456792 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.456814 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.559570 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.559613 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.559624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.559642 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.559654 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.567792 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/1.log" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.570252 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.570990 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.611319 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.622997 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.636363 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.651200 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.662292 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.662352 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.662363 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.662380 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.662392 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.664370 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.679017 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.690724 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.710746 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.721779 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.732282 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.744495 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.756369 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.764885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.764931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.764944 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.764962 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.764974 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.766409 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.780661 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.791572 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.801453 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.812923 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.827880 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:12Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.867229 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.867262 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.867273 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.867288 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.867299 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.969865 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.969914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.969926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.969943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:12 crc kubenswrapper[4727]: I0929 10:23:12.969954 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:12Z","lastTransitionTime":"2025-09-29T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.072265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.072306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.072314 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.072330 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.072356 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.108074 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:13 crc kubenswrapper[4727]: E0929 10:23:13.108650 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.108760 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:13 crc kubenswrapper[4727]: E0929 10:23:13.108951 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.108413 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:13 crc kubenswrapper[4727]: E0929 10:23:13.112038 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.174417 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.174637 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.174699 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.174783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.174854 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.276817 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.276858 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.276868 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.276883 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.276893 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.380684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.380735 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.380747 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.380766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.380778 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.483487 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.483766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.483875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.483962 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.484042 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.575208 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/2.log" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.575723 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/1.log" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.577686 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" exitCode=1 Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.577725 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.577756 4727 scope.go:117] "RemoveContainer" containerID="ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.578398 4727 scope.go:117] "RemoveContainer" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" Sep 29 10:23:13 crc kubenswrapper[4727]: E0929 10:23:13.578528 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.602050 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.602111 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.602130 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.602159 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.602172 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.604208 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.619204 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.633105 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.656089 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce41cb55bc5f91394bbe9fee19596614475ca1c5427e98fd433e02ea9e833260\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:22:58Z\\\",\\\"message\\\":\\\"{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI0929 10:22:58.509264 6259 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0929 10:22:58.509198 6259 services_controller.go:443] Built service openshift-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.58\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:22:58.509311 6259 services_controller.go:444] Built service openshift-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0929 10:22:58.509315 6259 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin netw\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.667249 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.682829 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.701697 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.704227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.704298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.704314 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.704351 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.704363 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.720229 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.734297 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.749390 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.765402 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.780832 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.793273 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.803208 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.806728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.806852 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.806939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.807045 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.807127 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.816921 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.830018 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.843566 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.857895 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.909876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.909915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.909927 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.909943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:13 crc kubenswrapper[4727]: I0929 10:23:13.909955 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:13Z","lastTransitionTime":"2025-09-29T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.012194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.012242 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.012257 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.012274 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.012287 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.107505 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.107661 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.114527 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.114572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.114582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.114597 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.114606 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.216823 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.216877 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.216889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.216909 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.216923 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.320095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.320137 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.320149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.320165 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.320193 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.443329 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.443629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.443713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.443805 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.443896 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.547250 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.547286 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.547295 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.547308 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.547319 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.548972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.549035 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.549046 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.549061 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.549071 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.562052 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.566096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.566143 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.566188 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.566207 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.566220 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.579754 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583018 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/2.log" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583794 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583850 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583864 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.583873 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.587128 4727 scope.go:117] "RemoveContainer" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.587553 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.597699 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.602695 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.602823 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.603148 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.603173 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.603188 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.603197 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.615718 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.621651 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.622258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.622461 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.622741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.622929 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.622953 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.636419 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.637104 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: E0929 10:23:14.637392 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.649115 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.650322 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.650390 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.650401 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.650413 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.650422 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.667246 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.678546 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.690520 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.703406 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.722804 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.737245 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.749459 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.754137 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.754169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.754182 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.754199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.754210 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.767764 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.778626 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.790234 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.805958 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.824352 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.836066 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.854032 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:14Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.856517 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.856637 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.856735 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.856816 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.856892 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.961201 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.961752 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.961873 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.962100 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:14 crc kubenswrapper[4727]: I0929 10:23:14.962223 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:14Z","lastTransitionTime":"2025-09-29T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.064405 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.064666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.064829 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.064971 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.065124 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.108103 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:15 crc kubenswrapper[4727]: E0929 10:23:15.108795 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.108647 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:15 crc kubenswrapper[4727]: E0929 10:23:15.109067 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.108869 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:15 crc kubenswrapper[4727]: E0929 10:23:15.109278 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.167165 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.168027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.168144 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.168247 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.168399 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.270620 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.270943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.271064 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.271166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.271260 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.374128 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.374156 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.374163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.374177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.374186 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.476515 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.476775 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.476848 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.476914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.476971 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.579453 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.579499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.579510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.579526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.579536 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.682565 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.682606 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.682615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.682633 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.682642 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.785258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.785302 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.785362 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.785381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.785390 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.887866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.887911 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.887923 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.887951 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.887965 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.990628 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.990666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.990674 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.990689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:15 crc kubenswrapper[4727]: I0929 10:23:15.990699 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:15Z","lastTransitionTime":"2025-09-29T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.093227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.093270 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.093281 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.093364 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.093379 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.107594 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:16 crc kubenswrapper[4727]: E0929 10:23:16.107744 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.197260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.197317 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.197328 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.197571 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.197583 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.300437 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.300488 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.300503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.300522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.300540 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.403649 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.403702 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.403713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.403730 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.403743 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.505755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.505797 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.505807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.505822 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.505833 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.608666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.608710 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.608719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.608733 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.608742 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.710564 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.710598 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.710607 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.710621 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.710631 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.812765 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.812805 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.812818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.812835 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.812847 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.915095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.915126 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.915134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.915154 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:16 crc kubenswrapper[4727]: I0929 10:23:16.915162 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:16Z","lastTransitionTime":"2025-09-29T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.017903 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.017949 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.017960 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.017976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.017988 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.108256 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.108289 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.108378 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:17 crc kubenswrapper[4727]: E0929 10:23:17.108431 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:17 crc kubenswrapper[4727]: E0929 10:23:17.108606 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:17 crc kubenswrapper[4727]: E0929 10:23:17.108787 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.120019 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.120059 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.120069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.120084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.120093 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.223065 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.223608 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.223684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.223773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.223837 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.326225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.326256 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.326265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.326280 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.326290 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.428827 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.428860 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.428869 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.428883 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.428892 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.531416 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.531931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.532008 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.532078 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.532157 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.634477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.634513 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.634522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.634536 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.634546 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.736862 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.736899 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.736926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.736942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.736953 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.839105 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.839141 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.839149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.839163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.839173 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.941367 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.941414 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.941429 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.941458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:17 crc kubenswrapper[4727]: I0929 10:23:17.941470 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:17Z","lastTransitionTime":"2025-09-29T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.043388 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.043441 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.043451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.043466 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.043475 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.108391 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:18 crc kubenswrapper[4727]: E0929 10:23:18.108520 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.146366 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.146406 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.146417 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.146433 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.146443 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.248861 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.248913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.248925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.248944 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.248955 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.350933 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.350985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.350998 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.351017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.351029 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.454377 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.454426 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.454437 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.454451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.454460 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.556540 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.556586 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.556596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.556611 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.556626 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.658874 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.658915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.658926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.658943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.658953 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.761266 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.761318 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.761330 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.761366 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.761380 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.863463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.863509 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.863525 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.863544 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.863555 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.965396 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.965423 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.965432 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.965446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:18 crc kubenswrapper[4727]: I0929 10:23:18.965455 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:18Z","lastTransitionTime":"2025-09-29T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.068275 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.068362 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.068374 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.068394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.068408 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.108349 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.108460 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:19 crc kubenswrapper[4727]: E0929 10:23:19.108574 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.108420 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:19 crc kubenswrapper[4727]: E0929 10:23:19.108783 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:19 crc kubenswrapper[4727]: E0929 10:23:19.108908 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.171039 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.171355 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.171594 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.171706 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.171806 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.274093 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.274145 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.274156 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.274174 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.274185 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.376581 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.376628 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.376638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.376653 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.376662 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.478714 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.478758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.478772 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.478789 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.478803 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.580629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.580667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.580675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.580689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.580699 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.683654 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.683766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.683788 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.683809 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.683824 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.786184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.786224 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.786233 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.786247 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.786257 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.888921 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.888959 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.888970 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.888987 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.889014 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.991732 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.991990 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.992102 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.992183 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:19 crc kubenswrapper[4727]: I0929 10:23:19.992299 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:19Z","lastTransitionTime":"2025-09-29T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.095000 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.095044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.095056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.095072 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.095084 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.107801 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:20 crc kubenswrapper[4727]: E0929 10:23:20.108057 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.197606 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.197921 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.198037 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.198145 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.198242 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.300842 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.300875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.300885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.300901 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.300913 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.402893 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.402933 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.402945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.402961 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.402972 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.505486 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.505547 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.505563 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.505589 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.505607 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.607881 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.607939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.607952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.607971 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.607986 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.711029 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.711074 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.711091 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.711113 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.711130 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.813416 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.813480 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.813501 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.813530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.813552 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.916919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.916969 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.916978 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.916993 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:20 crc kubenswrapper[4727]: I0929 10:23:20.917003 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:20Z","lastTransitionTime":"2025-09-29T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.019710 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.019751 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.019760 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.019776 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.019786 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.107772 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.107799 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:21 crc kubenswrapper[4727]: E0929 10:23:21.107889 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:21 crc kubenswrapper[4727]: E0929 10:23:21.107955 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.108391 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:21 crc kubenswrapper[4727]: E0929 10:23:21.108722 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.121031 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.122463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.122492 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.122503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.122515 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.122525 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.139927 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.150535 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.162771 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.174758 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.187041 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.199812 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.214603 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.233523 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.233786 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.233853 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.233918 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.233975 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.250923 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.269104 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.288350 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.300461 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.315555 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.326880 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.337655 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.337702 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.337722 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.337741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.337754 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.339890 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.361933 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.374282 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.386879 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:21Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.439638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.439672 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.439684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.439700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.439711 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.541852 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.541910 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.541925 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.541948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.541964 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.644517 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.644557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.644569 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.644583 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.644593 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.746930 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.746975 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.746989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.747009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.747020 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.849405 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.849444 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.849454 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.849467 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.849475 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.951859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.951912 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.951939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.951964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:21 crc kubenswrapper[4727]: I0929 10:23:21.951982 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:21Z","lastTransitionTime":"2025-09-29T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.055434 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.055670 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.055738 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.055832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.055891 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.108290 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:22 crc kubenswrapper[4727]: E0929 10:23:22.108954 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.157599 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.157817 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.157904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.158027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.158134 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.260674 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.260735 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.260748 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.260765 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.260777 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.362989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.363033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.363046 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.363063 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.363075 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.464987 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.465028 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.465036 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.465055 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.465064 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.567294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.567608 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.567688 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.567763 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.567825 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.670664 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.670698 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.671258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.671285 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.671296 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.774120 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.774183 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.774201 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.774226 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.774244 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.877621 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.877676 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.877701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.877732 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.877751 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.981303 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.981358 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.981368 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.981383 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:22 crc kubenswrapper[4727]: I0929 10:23:22.981394 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:22Z","lastTransitionTime":"2025-09-29T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.083893 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.084162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.084223 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.084287 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.084361 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.107545 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.107644 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:23 crc kubenswrapper[4727]: E0929 10:23:23.107940 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:23 crc kubenswrapper[4727]: E0929 10:23:23.108048 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.107665 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:23 crc kubenswrapper[4727]: E0929 10:23:23.108263 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.186487 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.186755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.186841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.186858 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.186875 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.292164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.292440 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.292520 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.292602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.292673 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.395636 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.396176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.396288 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.396394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.396480 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.499050 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.499075 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.499082 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.499095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.499103 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.601859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.601926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.601948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.601976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.601999 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.703923 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.704187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.704267 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.704380 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.704465 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.807002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.807056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.807068 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.807086 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.807096 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.910077 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.910129 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.910141 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.910162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:23 crc kubenswrapper[4727]: I0929 10:23:23.910172 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:23Z","lastTransitionTime":"2025-09-29T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.012466 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.012507 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.012518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.012532 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.012543 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.108099 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:24 crc kubenswrapper[4727]: E0929 10:23:24.108241 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.115068 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.115133 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.115146 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.115166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.115514 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.218408 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.218445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.218455 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.218471 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.218481 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.320754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.320792 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.320806 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.320822 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.320834 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.423330 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.423374 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.423382 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.423396 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.423406 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.525737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.525769 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.525777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.525789 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.525800 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.605031 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:24 crc kubenswrapper[4727]: E0929 10:23:24.605206 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:24 crc kubenswrapper[4727]: E0929 10:23:24.605463 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:23:56.605442126 +0000 UTC m=+106.778755488 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.628530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.628590 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.628610 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.628633 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.628649 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.730587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.730629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.730640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.730654 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.730663 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.833170 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.833557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.833677 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.833801 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.833916 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.936177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.936226 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.936238 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.936257 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:24 crc kubenswrapper[4727]: I0929 10:23:24.936269 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:24Z","lastTransitionTime":"2025-09-29T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.033737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.034180 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.034453 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.034692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.035025 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.049058 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:25Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.053722 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.053772 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.053788 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.053811 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.053829 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.066813 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:25Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.071109 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.071138 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.071148 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.071161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.071202 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.086076 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:25Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.090149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.090269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.090380 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.090503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.090574 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.104501 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:25Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.107441 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.107552 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.107605 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.107805 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.107887 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.107954 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.109064 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.109109 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.109126 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.109176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.109186 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.122841 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:25Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:25 crc kubenswrapper[4727]: E0929 10:23:25.123458 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.125513 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.125565 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.125574 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.125591 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.125600 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.227964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.228022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.228039 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.228061 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.228074 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.330741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.331103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.331191 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.331293 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.331359 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.433890 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.433938 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.433949 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.433970 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.433981 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.536414 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.536445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.536454 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.536469 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.536478 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.639960 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.640002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.640013 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.640027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.640037 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.742942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.743001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.743019 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.743037 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.743048 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.845811 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.845848 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.845860 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.845879 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.845890 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.949067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.949097 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.949107 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.949122 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:25 crc kubenswrapper[4727]: I0929 10:23:25.949134 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:25Z","lastTransitionTime":"2025-09-29T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.051271 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.051303 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.051311 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.051324 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.051333 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.107911 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:26 crc kubenswrapper[4727]: E0929 10:23:26.108189 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.154187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.154289 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.154311 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.154329 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.154409 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.256625 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.256665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.256674 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.256690 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.256702 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.359083 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.359128 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.359139 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.359156 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.359166 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.461391 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.461464 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.461475 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.461490 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.461499 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.564251 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.564399 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.564421 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.564480 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.564502 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.666607 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.666692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.666713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.666738 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.666791 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.768737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.768778 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.768787 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.768819 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.768830 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.871706 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.871738 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.871747 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.871760 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.871770 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.974660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.974710 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.974726 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.974743 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:26 crc kubenswrapper[4727]: I0929 10:23:26.974755 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:26Z","lastTransitionTime":"2025-09-29T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.077719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.077754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.077765 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.077780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.077791 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.107451 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.107450 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.107678 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:27 crc kubenswrapper[4727]: E0929 10:23:27.107707 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:27 crc kubenswrapper[4727]: E0929 10:23:27.107909 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:27 crc kubenswrapper[4727]: E0929 10:23:27.108125 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.109023 4727 scope.go:117] "RemoveContainer" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" Sep 29 10:23:27 crc kubenswrapper[4727]: E0929 10:23:27.109245 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.181874 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.181957 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.181974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.182007 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.182027 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.285671 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.285978 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.286105 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.286266 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.286623 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.389162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.389597 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.389678 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.389742 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.389812 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.492442 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.492478 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.492494 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.492518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.492531 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.595200 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.595265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.595285 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.595308 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.595321 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.697312 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.697371 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.697387 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.697405 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.697420 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.799208 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.799543 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.799640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.799749 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.799838 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.903205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.903241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.903252 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.903269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:27 crc kubenswrapper[4727]: I0929 10:23:27.903279 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:27Z","lastTransitionTime":"2025-09-29T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.005634 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.005671 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.005679 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.005694 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.005704 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107296 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:28 crc kubenswrapper[4727]: E0929 10:23:28.107475 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107583 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107594 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107607 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.107617 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.210534 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.210583 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.210601 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.210617 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.210628 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.313121 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.313158 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.313167 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.313183 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.313193 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.415242 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.415275 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.415284 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.415298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.415307 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.517621 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.517695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.517716 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.517742 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.517759 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.621101 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.621176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.621199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.621226 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.621248 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.723943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.723976 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.723985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.724018 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.724029 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.826515 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.826555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.826563 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.826579 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.826588 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.929160 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.929514 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.929624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.929713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:28 crc kubenswrapper[4727]: I0929 10:23:28.929799 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:28Z","lastTransitionTime":"2025-09-29T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.031863 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.031938 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.031950 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.031967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.031977 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.108264 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:29 crc kubenswrapper[4727]: E0929 10:23:29.108447 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.108518 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.108275 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:29 crc kubenswrapper[4727]: E0929 10:23:29.108661 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:29 crc kubenswrapper[4727]: E0929 10:23:29.108691 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.134245 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.134491 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.134615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.134701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.134796 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.237651 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.237691 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.237701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.237715 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.237724 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.340484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.340545 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.340557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.340573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.340589 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.443051 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.443106 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.443122 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.443142 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.443159 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.545078 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.545209 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.545270 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.545708 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.545745 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.648698 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.648742 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.648754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.648771 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.648784 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.751474 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.751510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.751521 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.751535 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.751547 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.853276 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.853316 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.853328 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.853363 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.853375 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.956087 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.956159 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.956169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.956195 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:29 crc kubenswrapper[4727]: I0929 10:23:29.956205 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:29Z","lastTransitionTime":"2025-09-29T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.059004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.059067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.059084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.059109 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.059126 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.107780 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:30 crc kubenswrapper[4727]: E0929 10:23:30.107928 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.162028 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.162102 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.162125 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.162158 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.162179 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.265030 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.265620 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.265969 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.266240 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.266918 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.369663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.369715 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.369730 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.369753 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.369768 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.472159 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.472264 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.472277 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.472294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.472309 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.575054 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.575131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.575149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.575175 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.575193 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.678243 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.678784 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.678995 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.679211 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.679455 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.782587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.782650 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.782665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.782686 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.782702 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.885423 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.885513 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.885539 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.885573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.885598 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.988622 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.988659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.988667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.988684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:30 crc kubenswrapper[4727]: I0929 10:23:30.988694 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:30Z","lastTransitionTime":"2025-09-29T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.092110 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.092178 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.092198 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.092237 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.092257 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.107672 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.107739 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.107768 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:31 crc kubenswrapper[4727]: E0929 10:23:31.107867 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:31 crc kubenswrapper[4727]: E0929 10:23:31.108046 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:31 crc kubenswrapper[4727]: E0929 10:23:31.108215 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.152584 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.167908 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.185873 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.196624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.196653 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.196663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.196680 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.196691 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.202383 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.220986 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.240227 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.259479 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.275414 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.293293 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.299216 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.299270 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.299287 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.299311 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.299326 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.322639 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.340070 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.358215 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.372703 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.388782 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.401099 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.401202 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.401213 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.401231 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.401242 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.407986 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.425108 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.441452 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.459471 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:31Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.504097 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.504153 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.504169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.504193 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.504205 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.606535 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.606574 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.606585 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.606779 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.606834 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.709717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.710011 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.710096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.710177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.710252 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.812989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.813031 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.813042 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.813058 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.813072 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.914833 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.915069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.915139 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.915208 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:31 crc kubenswrapper[4727]: I0929 10:23:31.915296 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:31Z","lastTransitionTime":"2025-09-29T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.017472 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.017550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.017562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.017582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.017592 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.108011 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:32 crc kubenswrapper[4727]: E0929 10:23:32.108132 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.119425 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.119485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.119503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.119520 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.119552 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.221962 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.222007 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.222018 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.222034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.222046 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.323841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.323919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.323948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.323977 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.323998 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.427024 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.427058 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.427069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.427085 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.427097 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.529635 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.529670 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.529679 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.529692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.529703 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.632457 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.633381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.633609 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.633799 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.633942 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.642120 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/0.log" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.642179 4727 generic.go:334] "Generic (PLEG): container finished" podID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" containerID="586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259" exitCode=1 Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.642214 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerDied","Data":"586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.642633 4727 scope.go:117] "RemoveContainer" containerID="586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.668767 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.690414 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.707611 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.729195 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.737571 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.737628 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.737640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.737655 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.737681 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.747153 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.767466 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.783543 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.806376 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.820363 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.834868 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.840698 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.840745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.840755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.840784 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.840802 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.847505 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.864560 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.877653 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.899142 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.916520 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.926929 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.937365 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.948243 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.948406 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.948420 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.948439 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.948460 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:32Z","lastTransitionTime":"2025-09-29T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:32 crc kubenswrapper[4727]: I0929 10:23:32.953808 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:32Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.051076 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.051114 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.051124 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.051162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.051176 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.107958 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.108062 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.108311 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:33 crc kubenswrapper[4727]: E0929 10:23:33.108117 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:33 crc kubenswrapper[4727]: E0929 10:23:33.108387 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:33 crc kubenswrapper[4727]: E0929 10:23:33.108605 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.153397 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.153435 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.153446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.153463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.153474 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.255876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.255951 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.255966 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.255989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.256003 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.359719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.359825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.359850 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.359889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.359912 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.463794 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.463858 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.463876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.463903 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.463921 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.567510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.567586 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.567602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.567629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.567647 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.647985 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/0.log" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.648078 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerStarted","Data":"e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.670122 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.670202 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.670228 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.670258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.670280 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.683459 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.698881 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.715620 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.742615 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.756991 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774024 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774286 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774319 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.774347 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.792226 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.808305 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.826133 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.843116 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.868196 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.877621 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.877681 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.877694 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.877716 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.877731 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.882829 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.897052 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.909127 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.923426 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.937625 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.950455 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.962684 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:33Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.980848 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.980927 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.980943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.980969 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:33 crc kubenswrapper[4727]: I0929 10:23:33.980983 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:33Z","lastTransitionTime":"2025-09-29T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.085187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.085309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.085333 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.085390 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.085414 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.111911 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:34 crc kubenswrapper[4727]: E0929 10:23:34.112139 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.189147 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.189641 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.189654 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.189678 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.189690 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.293744 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.293815 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.293828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.293862 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.293875 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.396649 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.396695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.396708 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.396728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.396738 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.499286 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.499386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.499402 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.499418 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.499428 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.608314 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.608383 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.608395 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.608411 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.608422 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.712242 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.712295 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.712303 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.712321 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.712330 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.815824 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.815889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.815905 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.815935 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.815951 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.920638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.920707 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.920719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.920743 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:34 crc kubenswrapper[4727]: I0929 10:23:34.920757 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:34Z","lastTransitionTime":"2025-09-29T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.024798 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.024867 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.024881 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.024905 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.024922 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.108139 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.108203 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.108192 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.108453 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.108606 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.108733 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.127605 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.127672 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.127709 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.127748 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.127776 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.230892 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.230950 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.230967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.230991 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.231008 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.334907 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.334967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.334980 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.335000 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.335017 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.438204 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.438283 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.438301 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.438366 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.438420 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.507818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.507891 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.507902 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.507922 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.507938 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.528057 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:35Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.535305 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.535434 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.535462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.535502 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.535529 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.551517 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:35Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.556568 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.556612 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.556624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.556645 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.556658 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.578381 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:35Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.583151 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.583205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.583227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.583259 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.583280 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.600945 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:35Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.606605 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.606667 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.606682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.606705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.606722 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.622925 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:35Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:35 crc kubenswrapper[4727]: E0929 10:23:35.623110 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.625044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.625112 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.625128 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.625151 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.625165 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.728969 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.729063 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.729078 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.729103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.729126 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.833533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.833617 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.833643 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.833671 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.833693 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.937424 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.937492 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.937511 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.937541 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:35 crc kubenswrapper[4727]: I0929 10:23:35.937562 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:35Z","lastTransitionTime":"2025-09-29T10:23:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.040869 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.040926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.040936 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.040952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.040964 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.107828 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:36 crc kubenswrapper[4727]: E0929 10:23:36.108079 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.143656 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.143690 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.143701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.143719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.143736 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.247259 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.247327 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.247365 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.247389 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.247403 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.350422 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.350521 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.350545 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.350582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.350611 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.454162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.454245 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.454269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.454306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.454401 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.557304 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.557416 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.557432 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.557455 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.557468 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.668049 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.668094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.668106 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.668126 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.668138 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.771883 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.771936 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.771980 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.772009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.772026 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.875739 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.875808 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.875822 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.875857 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.875878 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.979711 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.979773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.979788 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.979810 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:36 crc kubenswrapper[4727]: I0929 10:23:36.979822 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:36Z","lastTransitionTime":"2025-09-29T10:23:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.083816 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.083856 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.083865 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.083882 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.083894 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.107522 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.107669 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:37 crc kubenswrapper[4727]: E0929 10:23:37.107738 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:37 crc kubenswrapper[4727]: E0929 10:23:37.107874 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.107672 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:37 crc kubenswrapper[4727]: E0929 10:23:37.108140 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.187161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.187201 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.187209 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.187225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.187234 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.290606 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.290663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.290683 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.290711 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.290732 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.394286 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.394352 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.394364 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.394380 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.394390 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.496447 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.496550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.496573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.496602 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.496622 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.599859 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.599919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.599943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.599972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.599994 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.702745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.702803 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.702820 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.702844 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.702861 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.806148 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.806215 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.806231 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.806254 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.806269 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.909916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.909981 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.910001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.910025 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:37 crc kubenswrapper[4727]: I0929 10:23:37.910042 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:37Z","lastTransitionTime":"2025-09-29T10:23:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.013458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.013519 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.013532 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.013555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.013567 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.107942 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:38 crc kubenswrapper[4727]: E0929 10:23:38.108292 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.116257 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.116345 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.116356 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.116381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.116391 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.220125 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.220173 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.220184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.220201 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.220211 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.322989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.323042 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.323055 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.323073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.323085 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.426224 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.426301 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.426315 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.426374 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.426390 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.529153 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.529215 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.529228 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.529251 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.529265 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.632665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.632705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.632714 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.632727 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.632736 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.735381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.735425 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.735442 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.735467 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.735484 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.838038 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.838174 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.838194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.838221 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.838238 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.940479 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.940518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.940526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.940540 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:38 crc kubenswrapper[4727]: I0929 10:23:38.940549 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:38Z","lastTransitionTime":"2025-09-29T10:23:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.043445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.043611 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.043634 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.043662 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.043682 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.108308 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:39 crc kubenswrapper[4727]: E0929 10:23:39.108520 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.108653 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.108780 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:39 crc kubenswrapper[4727]: E0929 10:23:39.108921 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:39 crc kubenswrapper[4727]: E0929 10:23:39.109178 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.126066 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.147834 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.147904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.147914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.147947 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.147960 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.251328 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.251421 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.251477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.251505 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.251523 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.354298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.354401 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.354423 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.354456 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.354481 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.457619 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.457681 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.457694 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.457713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.457727 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.560648 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.560724 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.560745 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.560777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.560798 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.664150 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.664187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.664198 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.664215 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.664228 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.767415 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.767465 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.767492 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.767514 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.767527 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.870160 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.870237 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.870257 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.870284 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.870304 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.974264 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.974313 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.974325 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.974360 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:39 crc kubenswrapper[4727]: I0929 10:23:39.974373 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:39Z","lastTransitionTime":"2025-09-29T10:23:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.076923 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.076964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.076975 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.076989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.077000 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.107906 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:40 crc kubenswrapper[4727]: E0929 10:23:40.108168 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.179117 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.179166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.179174 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.179188 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.179197 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.281988 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.282032 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.282042 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.282056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.282064 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.384251 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.384309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.384323 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.384365 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.384383 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.486271 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.486302 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.486309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.486323 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.486356 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.588149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.588184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.588192 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.588205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.588214 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.691117 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.691154 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.691163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.691177 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.691188 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.793136 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.793176 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.793187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.793205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.793216 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.895967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.896023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.896033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.896048 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.896056 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:40Z","lastTransitionTime":"2025-09-29T10:23:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:40 crc kubenswrapper[4727]: I0929 10:23:40.913566 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:40 crc kubenswrapper[4727]: E0929 10:23:40.913715 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:23:40 crc kubenswrapper[4727]: E0929 10:23:40.913815 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:24:44.91379451 +0000 UTC m=+155.087107882 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.002660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.002743 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.002787 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.002819 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.002842 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.014157 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.014436 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.014528 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.014580 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014713 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:45.01467448 +0000 UTC m=+155.187987882 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014811 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014821 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014863 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014882 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014906 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:24:45.014877895 +0000 UTC m=+155.188191307 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014906 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014963 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014960 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:24:45.014934366 +0000 UTC m=+155.188247768 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.014984 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.015031 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:24:45.015016978 +0000 UTC m=+155.188330380 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.105639 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.105684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.105695 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.105713 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.105727 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.111487 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.111548 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.111665 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.111721 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.111891 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:41 crc kubenswrapper[4727]: E0929 10:23:41.112024 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.141142 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.155286 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.166784 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.176710 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.188501 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.202041 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.207174 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.207232 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.207243 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.207260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.207272 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.213849 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.225105 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.238743 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.258606 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.269971 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da604f80-d35e-4354-9b4c-912b35144858\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.282688 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.297279 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.310847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.310904 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.310929 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.310949 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.310961 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.311685 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.326956 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.341891 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.357430 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.374078 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.390837 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:41Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.413725 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.413809 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.413821 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.413835 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.413843 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.516466 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.516518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.516529 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.516547 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.516561 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.619227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.619370 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.619388 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.619414 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.619431 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.721754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.721825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.721840 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.721866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.721881 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.825143 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.825181 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.825189 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.825203 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.825211 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.926675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.926712 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.926724 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.926742 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:41 crc kubenswrapper[4727]: I0929 10:23:41.926755 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:41Z","lastTransitionTime":"2025-09-29T10:23:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.030362 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.030435 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.030450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.030481 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.030496 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.107952 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:42 crc kubenswrapper[4727]: E0929 10:23:42.108268 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.110135 4727 scope.go:117] "RemoveContainer" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.134061 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.134175 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.134192 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.134217 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.134237 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.236878 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.236917 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.236926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.236941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.236951 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.341682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.341782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.341805 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.341862 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.341879 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.446856 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.446944 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.446980 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.447024 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.447045 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.551371 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.551450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.551465 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.551490 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.551526 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.654663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.654733 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.654753 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.654782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.654802 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.679867 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/2.log" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.683199 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.683772 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.709313 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.741267 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.757992 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.758034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.758045 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.758063 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.758077 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.768191 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.784293 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.807383 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.821202 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.833118 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.847067 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da604f80-d35e-4354-9b4c-912b35144858\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.860772 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.861483 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.861530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.861542 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.861561 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.861573 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.875276 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.891017 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.909550 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.925469 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.938378 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.951477 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.962969 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.963709 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.963741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.963750 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.963771 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.963785 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:42Z","lastTransitionTime":"2025-09-29T10:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:42 crc kubenswrapper[4727]: I0929 10:23:42.982203 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.000205 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:42Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.012262 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.066397 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.066442 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.066463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.066481 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.066492 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.108598 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.108599 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.108739 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:43 crc kubenswrapper[4727]: E0929 10:23:43.108893 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:43 crc kubenswrapper[4727]: E0929 10:23:43.109058 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:43 crc kubenswrapper[4727]: E0929 10:23:43.109153 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.168377 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.168615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.168623 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.168638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.168648 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.271131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.271170 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.271194 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.271212 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.271224 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.374361 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.374398 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.374408 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.374440 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.374454 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.477216 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.477272 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.477282 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.477300 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.477310 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.580518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.580608 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.580629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.580660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.580682 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.695992 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.696020 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.696029 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.696042 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.696053 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.697607 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/3.log" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.698034 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/2.log" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.700381 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" exitCode=1 Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.700413 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.700487 4727 scope.go:117] "RemoveContainer" containerID="3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.700989 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:23:43 crc kubenswrapper[4727]: E0929 10:23:43.701113 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.714924 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.730769 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.745861 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.756828 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.775739 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.788800 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.799166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.799225 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.799236 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.799253 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.799264 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.801717 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.811207 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.826015 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.840545 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.853746 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.870955 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.901717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.902188 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.902502 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.902655 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.902804 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:43Z","lastTransitionTime":"2025-09-29T10:23:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.904157 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ccf4f82c71a86de3ef5ea310a017f249f74a13a404c245044e23f9e3f1c6ba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:13Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc\\\\nI0929 10:23:13.339869 6452 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0929 10:23:13.339892 6452 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:13Z is after 2025-08-24T17:21:41Z]\\\\nI0929 10:23:13.339892 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:43Z\\\",\\\"message\\\":\\\"21 services_controller.go:443] Built service openshift-console/console LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.194\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:23:43.145000 6821 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI0929 10:23:43.146332 6821 services_controller.go:444] Built service openshift-console/console LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146365 6821 services_controller.go:445] Built service openshift-console/console LB template configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146324 6821 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 10:23:43.146385 6821 services_controller.go:451] Buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.919734 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.935315 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da604f80-d35e-4354-9b4c-912b35144858\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.952266 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.967797 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.981622 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:43 crc kubenswrapper[4727]: I0929 10:23:43.997251 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:43Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.005034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.005061 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.005071 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.005090 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.005104 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107320 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:44 crc kubenswrapper[4727]: E0929 10:23:44.107483 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107623 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107707 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107734 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.107793 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.209997 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.210041 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.210053 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.210068 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.210080 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.312587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.312668 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.312687 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.312712 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.312729 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.415636 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.415690 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.415706 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.415728 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.415745 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.518635 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.518703 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.518729 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.518761 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.518783 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.622009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.622142 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.622166 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.622199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.622228 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.707779 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/3.log" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.713585 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:23:44 crc kubenswrapper[4727]: E0929 10:23:44.713874 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.725884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.725951 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.725983 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.726014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.726036 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.738274 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.760402 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.777472 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.805660 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.822821 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da604f80-d35e-4354-9b4c-912b35144858\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.831300 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.831387 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.831403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.831432 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.831450 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.841751 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.858268 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.877460 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.897811 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.918229 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.935095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.935149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.935162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.935182 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.935193 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:44Z","lastTransitionTime":"2025-09-29T10:23:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.939091 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.966954 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:44 crc kubenswrapper[4727]: I0929 10:23:44.982330 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:44Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.017136 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.038853 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.038896 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.038907 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.038928 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.038941 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.046444 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.080120 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:43Z\\\",\\\"message\\\":\\\"21 services_controller.go:443] Built service openshift-console/console LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.194\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:23:43.145000 6821 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI0929 10:23:43.146332 6821 services_controller.go:444] Built service openshift-console/console LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146365 6821 services_controller.go:445] Built service openshift-console/console LB template configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146324 6821 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 10:23:43.146385 6821 services_controller.go:451] Buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.098932 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.108125 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.108318 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.108691 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.108796 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.109027 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.109222 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.128456 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.142304 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.142400 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.142418 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.142446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.142462 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.148804 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.245181 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.245241 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.245265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.245295 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.245318 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.349162 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.349234 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.349260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.349299 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.349325 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.454047 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.454104 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.454123 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.454149 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.454168 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.558462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.558538 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.558558 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.558587 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.558613 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.662968 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.663034 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.663054 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.663085 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.663106 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.766729 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.766777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.766790 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.766807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.766819 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.870219 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.870255 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.870289 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.870306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.870316 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.930666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.930744 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.930761 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.930786 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.930800 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.954875 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.959693 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.959767 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.959784 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.959807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.959823 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.973942 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.977496 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.977556 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.977570 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.977586 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.977599 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:45 crc kubenswrapper[4727]: E0929 10:23:45.991950 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:45Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.996310 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.996394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.996412 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.996438 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:45 crc kubenswrapper[4727]: I0929 10:23:45.996457 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:45Z","lastTransitionTime":"2025-09-29T10:23:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: E0929 10:23:46.012861 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.017692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.017755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.017767 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.017811 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.017826 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: E0929 10:23:46.030782 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:46Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:46 crc kubenswrapper[4727]: E0929 10:23:46.031036 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.033692 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.033784 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.033804 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.033862 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.033882 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.107804 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:46 crc kubenswrapper[4727]: E0929 10:23:46.107972 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.137411 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.137467 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.137485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.137505 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.137519 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.240409 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.240445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.240627 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.240659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.240668 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.343411 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.343490 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.343508 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.343533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.343551 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.446072 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.446115 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.446126 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.446145 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.446156 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.548777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.548818 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.548828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.548846 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.548857 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.651455 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.651495 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.651505 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.651522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.651535 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.754553 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.754613 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.754638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.754666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.754687 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.859073 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.859199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.859285 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.859434 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.859454 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.965290 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.965370 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.965385 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.965403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:46 crc kubenswrapper[4727]: I0929 10:23:46.965412 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:46Z","lastTransitionTime":"2025-09-29T10:23:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.068542 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.068609 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.068629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.068705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.068736 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.108564 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.108585 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:47 crc kubenswrapper[4727]: E0929 10:23:47.108823 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.108594 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:47 crc kubenswrapper[4727]: E0929 10:23:47.108973 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:47 crc kubenswrapper[4727]: E0929 10:23:47.109037 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.170937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.170968 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.170977 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.170989 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.170999 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.277017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.277556 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.277698 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.277846 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.277957 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.381381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.381442 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.381453 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.381477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.381497 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.485055 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.485108 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.485118 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.485138 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.485150 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.589800 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.590171 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.590255 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.590427 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.590534 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.694923 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.694988 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.695003 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.695029 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.695040 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.870624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.870690 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.870707 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.870734 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.870750 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.974503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.974572 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.974582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.974604 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:47 crc kubenswrapper[4727]: I0929 10:23:47.974618 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:47Z","lastTransitionTime":"2025-09-29T10:23:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.078142 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.078232 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.078263 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.078294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.078313 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.107838 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:48 crc kubenswrapper[4727]: E0929 10:23:48.108030 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.182017 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.182069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.182082 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.182103 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.182117 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.285393 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.285461 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.285482 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.285508 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.285526 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.388463 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.388536 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.388554 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.388582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.388601 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.492258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.492321 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.492366 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.492390 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.492407 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.596485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.596546 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.596560 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.596583 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.596596 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.700482 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.700548 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.700565 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.700596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.700617 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.804775 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.804837 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.804860 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.804883 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.804896 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.908768 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.908840 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.908854 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.908876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:48 crc kubenswrapper[4727]: I0929 10:23:48.908899 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:48Z","lastTransitionTime":"2025-09-29T10:23:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.012995 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.013044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.013054 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.013077 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.013088 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.107553 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.107621 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.107698 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:49 crc kubenswrapper[4727]: E0929 10:23:49.107742 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:49 crc kubenswrapper[4727]: E0929 10:23:49.107926 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:49 crc kubenswrapper[4727]: E0929 10:23:49.108061 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.115181 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.115249 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.115267 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.115298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.115317 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.219052 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.219159 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.219207 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.219236 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.219252 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.322919 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.322988 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.323002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.323029 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.323044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.425909 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.425996 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.426023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.426060 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.426085 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.529038 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.529084 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.529094 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.529110 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.529120 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.632425 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.632485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.632499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.632527 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.632541 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.735675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.735788 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.735809 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.735831 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.735890 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.838827 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.838901 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.838913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.838939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.838955 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.942079 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.942132 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.942145 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.942163 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:49 crc kubenswrapper[4727]: I0929 10:23:49.942176 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:49Z","lastTransitionTime":"2025-09-29T10:23:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.044283 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.044389 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.044415 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.044445 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.044469 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.108308 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:50 crc kubenswrapper[4727]: E0929 10:23:50.108600 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.147603 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.147632 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.147651 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.147669 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.147680 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.250875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.250906 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.250914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.250927 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.250935 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.353635 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.353689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.353700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.353718 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.353729 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.458265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.458394 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.458424 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.458461 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.458497 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.561686 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.561780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.561806 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.561835 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.561856 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.664453 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.664526 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.664550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.664578 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.664594 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.766326 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.766385 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.766393 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.766407 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.766415 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.869405 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.869451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.869464 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.869485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.869501 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.972276 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.972421 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.972448 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.972481 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:50 crc kubenswrapper[4727]: I0929 10:23:50.972505 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:50Z","lastTransitionTime":"2025-09-29T10:23:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.075948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.076021 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.076045 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.076077 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.076094 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.108011 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:51 crc kubenswrapper[4727]: E0929 10:23:51.108256 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.108470 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:51 crc kubenswrapper[4727]: E0929 10:23:51.108626 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.108831 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:51 crc kubenswrapper[4727]: E0929 10:23:51.108919 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.129263 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc578ace-78c8-4d17-a556-c6d6ceb149a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:43Z\\\",\\\"message\\\":\\\"21 services_controller.go:443] Built service openshift-console/console LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.194\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 10:23:43.145000 6821 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI0929 10:23:43.146332 6821 services_controller.go:444] Built service openshift-console/console LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146365 6821 services_controller.go:445] Built service openshift-console/console LB template configs for network=default: []services.lbConfig(nil)\\\\nI0929 10:23:43.146324 6821 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0929 10:23:43.146385 6821 services_controller.go:451] Buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:23:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq6pj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lqql2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.139852 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6lmbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02f5f1cc-410d-4f33-bf6c-8a4f06830e58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b50fc589d04fc78bdeca00657e25a696ade811bd17b2e435e134c9141fbafad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slgjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:42Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6lmbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.155102 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gttbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59055eb3-6cad-4335-b100-39955f1c0500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btrmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gttbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.171131 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ea5aad8-864f-457a-830a-f86b3b8a555f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf4f506082f833ed55f9193cee9ea4dc6720481c0a3a48fdcd332d9a49cb1e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e222a594666c27231554f8807880fe14f35824e8f0c3ae815682cc82bb7d6ee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf110ab6f7e56d71d1be6597ee137d3c80bc01a35eccdcd8ccc5622749bbef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3038f2cbf40b16950c102c1337673cc04d08d668630f3e73ae667da032e2fe5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d48d6618250cd12256e347883b55ae3e44645d562a90580124b6c8b431d1d8aa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T10:22:30Z\\\",\\\"message\\\":\\\"W0929 10:22:18.864992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0929 10:22:18.865441 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759141338 cert, and key in /tmp/serving-cert-2990911016/serving-signer.crt, /tmp/serving-cert-2990911016/serving-signer.key\\\\nI0929 10:22:19.735773 1 observer_polling.go:159] Starting file observer\\\\nW0929 10:22:19.740551 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0929 10:22:19.740775 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 10:22:19.741771 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2990911016/tls.crt::/tmp/serving-cert-2990911016/tls.key\\\\\\\"\\\\nF0929 10:22:30.190011 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11253645aa6a5c977619b5f71992b7f27b5fc1fbc2aa01905510ceb01687249f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e651460d690e49180bfb1ae77a51005120154afec410ce091063c00b606c2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.178516 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.178571 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.178582 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.178598 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.178609 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.186594 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.199461 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.212000 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3bfd8eb3add9b121a0c286caf70c66a12280d28303b66e125c77a24d02e730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2677a0a22111f6dbcd6b8780549853c210b43cb607798b9e8e3538dfec367a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.229625 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c74616e3-ccb3-453b-a7b7-b734d3203c06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecaa6f13a4275c414f1b4f689a1c6ff0a1c2229f2042e3886ad8bc01b7f18c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94334900da095639c0e9641602f25d956fa97d22ef7c41ca9f2e03c2c2f71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4101e5a5e07cac7e6b4dfaf90a77a8f4a7a526b27748328e1a6ce0f02f26c2b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a0f57cc576375da238ded41ee8004f9c46de3c4f0ea4196f6f96eb13b901f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa21979bf8aa798f99fd3e894d7f0079b3d6bc682b68d27bcc6872000949e5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6259a5fb75ae095752376c3e5bbb4c6d0910107e8037b9179d97e087353a45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caeaf1a7fdc2dd0929700e2d8e7f5de494038998d4722af1c73ae24b75f9f72e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsb2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vg7wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.240753 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da604f80-d35e-4354-9b4c-912b35144858\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed2b093255677fb89c9beee610407e6443d31e93282bad074c38f4bcfdc3f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://793342615c1fb32e870a542a9984bc455e45c0de289263b1ad5a74bb851ad90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.253452 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7244fafa-843a-4162-8dd4-6700102c1c56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85c07869649027c25709c1713ff4f9290cf0c0f4ee02e56b01b6791d5642c562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c64f70ddf3acd1f0f9054f08d5050e1f05d1918231455e7d80b8394416c832d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1c818e86036a535b2158affe2906f9978affee353a2df712a7ab418a20aa6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://572182ef7b140cda312164c50eb95b75f7c3fc06f39d8de86c728f6f9b2a270d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.267462 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.280743 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mvx5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd0ac62e-15bc-4508-a7da-414196360165\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f1e996f36d714034ad65d573172e7c6afab8e4af5e04e09aa849138851a7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c6vlv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mvx5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.281808 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.281852 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.281866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.281886 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.281900 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.293817 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea96a49842fa0da17967d1761f5fb6001b251c8e3c277fe51c7ae856152f8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.304316 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26551cb898e1fa7080e70f78e3e5f4e81c14b12ed82c2e740ec8a25043f15f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.316005 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwdcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e3b98d9-a9fc-4bf3-8053-b7701c047d99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T10:23:31Z\\\",\\\"message\\\":\\\"2025-09-29T10:22:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2\\\\n2025-09-29T10:22:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92139248-1f0d-4001-a9e2-68db270e9aa2 to /host/opt/cni/bin/\\\\n2025-09-29T10:22:46Z [verbose] multus-daemon started\\\\n2025-09-29T10:22:46Z [verbose] Readiness Indicator file check\\\\n2025-09-29T10:23:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:23:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44csc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwdcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.351632 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cdf5dfa-ea0f-48e8-9f56-df22343f7d58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://136dc7fdc241ebfc507f0897f57942bb5887dfa805457317e01b199ad713e517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb9fcf98ddb180b471daf1347fcb131070c0dc038ffdebfb6ed014fdd086cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfmmt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-d8zdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.371439 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b87b601-efaa-49b5-8dd5-f95f567b7207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98e40c5aabf719dad6b33bdaafdce17757828cb80b44bee0bbe7867d56628a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caf5d277a87cec365c370cba59ddde878737a37e38ca6d4c44840c94a88c87bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://041b58b2ae7c34c6f06c84b4eeaab79ba3fa697c34af66202377ac79ae534e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71d0d1e35d2238962788a57709580cae6afef33ec7f7f21abe349d5e9a0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30140e5f924a5d6360a319c3a84eef88d15e4f9230301532184e6477a32b0e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ed5b8da902636c36c56feaa862317846566b14b99b75bf2688863b385dde14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab7ca5866525fdec6d0b8a6ca5303475bbf9bbd58b47f72fcb49d881b50ba3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf786b0abfd444e8175ce0f07e92b912604041f702e1faa64fb2a4d482dc336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T10:22:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.383711 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d883ae71-80d1-430c-9afd-b5970c67ae6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04405bd6a4cc56e0dced0d04b8b0431f81fbf37d6087bfb15e075b3b8871fca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78bec2f291414a1f31f9d1b02ee16c830e4f425e7c166a57bbead46f21bcb6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a014f90fdbd6b052718b9ac2e92ed627b16c5341ca50a5928142ebfda7bfd8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://195e948f12be882be7ebca136cd875f6746671fc99b55f2ea964c2b5ac08d42c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.388054 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.388125 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.388134 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.388169 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.388182 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.396188 4727 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791106af-2f47-4c80-9f20-e0bc4131f833\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T10:22:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b33d030aa92c6f66531e2e88851be1755bb0d3a247ba68128074b32856836603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T10:22:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fj5ch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T10:22:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w25jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:51Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.490429 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.490485 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.490493 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.490507 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.490515 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.593773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.594115 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.594132 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.594184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.594204 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.697426 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.697478 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.697491 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.697508 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.697520 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.800006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.800051 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.800062 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.800079 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.800092 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.902063 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.902107 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.902117 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.902132 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:51 crc kubenswrapper[4727]: I0929 10:23:51.902142 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:51Z","lastTransitionTime":"2025-09-29T10:23:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.004399 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.004431 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.004441 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.004455 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.004466 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.106499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.106561 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.106588 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.106619 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.106641 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.107477 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:52 crc kubenswrapper[4727]: E0929 10:23:52.107678 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.210171 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.210204 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.210219 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.210235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.210245 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.311841 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.311875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.311883 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.311896 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.311904 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.414294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.414353 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.414364 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.414378 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.414386 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.516825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.516860 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.516871 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.516886 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.516896 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.619322 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.619372 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.619381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.619397 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.619407 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.721590 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.721646 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.721662 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.721683 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.721697 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.823956 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.824006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.824022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.824043 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.824056 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.926701 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.926741 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.926751 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.926766 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:52 crc kubenswrapper[4727]: I0929 10:23:52.926777 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:52Z","lastTransitionTime":"2025-09-29T10:23:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.029168 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.029228 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.029240 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.029258 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.029271 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.108482 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.108498 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.108506 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:53 crc kubenswrapper[4727]: E0929 10:23:53.108839 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:53 crc kubenswrapper[4727]: E0929 10:23:53.108683 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:53 crc kubenswrapper[4727]: E0929 10:23:53.109005 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.131586 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.131678 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.131922 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.131943 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.131958 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.234477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.234522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.234533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.234550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.234561 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.337153 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.337261 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.337281 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.337373 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.337399 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.440214 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.440252 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.440261 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.440274 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.440284 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.543490 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.543528 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.543537 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.543550 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.543560 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.647376 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.647456 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.647477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.647511 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.647532 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.750120 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.750186 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.750195 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.750214 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.750224 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.853158 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.853240 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.853260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.853294 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.853319 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.955686 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.955725 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.955733 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.955748 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:53 crc kubenswrapper[4727]: I0929 10:23:53.955758 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:53Z","lastTransitionTime":"2025-09-29T10:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.058678 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.058737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.058753 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.058772 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.058787 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.107556 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:54 crc kubenswrapper[4727]: E0929 10:23:54.107713 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.161937 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.161991 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.162006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.162030 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.162050 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.264438 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.264491 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.264503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.264520 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.264536 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.367552 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.367607 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.367615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.367630 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.367639 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.470500 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.470545 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.470557 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.470577 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.470589 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.573320 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.573392 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.573403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.573419 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.573430 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.676238 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.676308 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.676319 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.676344 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.676388 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.777830 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.777887 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.777899 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.777915 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.777925 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.880503 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.880545 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.880555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.880575 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.880588 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.982902 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.982970 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.982982 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.983008 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:54 crc kubenswrapper[4727]: I0929 10:23:54.983017 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:54Z","lastTransitionTime":"2025-09-29T10:23:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.085136 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.085180 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.085191 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.085207 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.085220 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.108072 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.108124 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.108146 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:55 crc kubenswrapper[4727]: E0929 10:23:55.108306 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:55 crc kubenswrapper[4727]: E0929 10:23:55.108483 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:55 crc kubenswrapper[4727]: E0929 10:23:55.108617 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.109291 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:23:55 crc kubenswrapper[4727]: E0929 10:23:55.109598 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.187638 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.187694 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.187703 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.187717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.187729 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.290780 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.290822 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.290832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.290847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.290856 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.393624 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.393684 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.393700 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.393722 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.393738 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.497542 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.497605 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.497629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.497660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.497700 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.600292 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.600376 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.600393 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.600414 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.600431 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.703205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.703242 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.703254 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.703271 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.703282 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.805980 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.806033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.806044 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.806063 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.806082 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.908433 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.908494 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.908510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.908534 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:55 crc kubenswrapper[4727]: I0929 10:23:55.908551 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:55Z","lastTransitionTime":"2025-09-29T10:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.012032 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.012097 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.012116 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.012141 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.012162 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.107605 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.107749 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.116227 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.116298 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.116311 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.116381 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.116399 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.218566 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.218612 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.218661 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.218679 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.218691 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.316838 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.316909 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.316931 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.316961 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.316984 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.329180 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:56Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.333314 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.333389 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.333398 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.333427 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.333436 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.348266 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:56Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.352720 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.352765 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.352782 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.352802 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.352817 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.368036 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:56Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.371942 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.371972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.371986 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.372004 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.372016 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.384133 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:56Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.388161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.388218 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.388236 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.388259 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.388281 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.416780 4727 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T10:23:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4ffff91-59e9-4746-8297-f21cac254b29\\\",\\\"systemUUID\\\":\\\"833ae235-acd8-410a-9184-a9bb33bace31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T10:23:56Z is after 2025-08-24T17:21:41Z" Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.416992 4727 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.423827 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.423876 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.423892 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.423916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.423933 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.527080 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.527164 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.527181 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.527212 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.527272 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.618806 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.618978 4727 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:56 crc kubenswrapper[4727]: E0929 10:23:56.619043 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs podName:59055eb3-6cad-4335-b100-39955f1c0500 nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.619027948 +0000 UTC m=+170.792341310 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs") pod "network-metrics-daemon-gttbj" (UID: "59055eb3-6cad-4335-b100-39955f1c0500") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.629106 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.629136 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.629145 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.629157 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.629166 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.732199 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.732240 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.732250 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.732268 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.732277 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.834615 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.834685 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.834703 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.834726 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.834742 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.936985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.937036 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.937048 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.937065 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:56 crc kubenswrapper[4727]: I0929 10:23:56.937080 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:56Z","lastTransitionTime":"2025-09-29T10:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.039511 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.039543 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.039551 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.039564 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.039574 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.108001 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.108117 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:57 crc kubenswrapper[4727]: E0929 10:23:57.108123 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:57 crc kubenswrapper[4727]: E0929 10:23:57.108214 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.108251 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:57 crc kubenswrapper[4727]: E0929 10:23:57.108303 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.141689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.141716 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.141725 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.141737 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.141747 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.243755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.243783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.243791 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.243805 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.243813 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.346012 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.346046 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.346056 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.346069 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.346079 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.448403 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.448450 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.448461 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.448477 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.448487 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.550522 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.550554 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.550562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.550575 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.550585 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.652835 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.652875 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.652886 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.652899 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.652911 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.756664 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.756721 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.756736 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.756754 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.756766 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.859885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.859940 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.859953 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.859974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.859993 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.962398 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.962439 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.962451 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.962468 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:57 crc kubenswrapper[4727]: I0929 10:23:57.962480 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:57Z","lastTransitionTime":"2025-09-29T10:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.064940 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.064988 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.065003 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.065022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.065034 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.107495 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:23:58 crc kubenswrapper[4727]: E0929 10:23:58.107678 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.167631 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.167675 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.167686 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.167704 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.167715 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.269680 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.269755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.269773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.269794 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.269805 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.372433 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.372473 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.372484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.372502 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.372532 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.474589 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.474633 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.474643 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.474660 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.474671 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.576874 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.576949 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.576972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.577001 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.577023 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.679719 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.679778 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.679800 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.679820 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.679833 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.783192 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.783238 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.783261 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.783280 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.783293 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.885753 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.885807 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.885828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.885849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.885862 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.988193 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.988244 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.988256 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.988273 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:58 crc kubenswrapper[4727]: I0929 10:23:58.988285 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:58Z","lastTransitionTime":"2025-09-29T10:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.090462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.090506 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.090515 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.090532 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.090541 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.107635 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.107783 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:23:59 crc kubenswrapper[4727]: E0929 10:23:59.108098 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.108178 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:23:59 crc kubenswrapper[4727]: E0929 10:23:59.108323 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:23:59 crc kubenswrapper[4727]: E0929 10:23:59.108431 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.193113 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.193161 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.193173 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.193187 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.193197 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.296150 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.296224 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.296235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.296250 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.296259 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.398640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.398679 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.398689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.398705 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.398715 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.500954 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.501002 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.501014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.501032 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.501044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.603872 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.603913 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.603923 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.603941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.603953 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.707555 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.707600 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.707609 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.707623 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.707636 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.809845 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.809916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.809935 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.809959 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.809977 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.912781 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.912836 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.912848 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.912866 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:23:59 crc kubenswrapper[4727]: I0929 10:23:59.912877 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:23:59Z","lastTransitionTime":"2025-09-29T10:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.015462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.015533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.015553 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.015577 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.015598 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.108323 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:00 crc kubenswrapper[4727]: E0929 10:24:00.108855 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.117912 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.117948 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.117974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.117988 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.117999 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.220697 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.220758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.220781 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.220805 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.220825 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.323435 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.323484 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.323495 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.323516 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.323532 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.426278 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.426825 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.426900 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.426978 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.427047 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.529961 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.529996 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.530006 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.530023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.530035 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.632596 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.632873 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.632966 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.633061 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.633145 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.735609 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.735659 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.735670 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.735690 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.735704 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.838589 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.838641 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.838661 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.838689 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.838712 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.942504 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.942559 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.942581 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.942607 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:00 crc kubenswrapper[4727]: I0929 10:24:00.942627 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:00Z","lastTransitionTime":"2025-09-29T10:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.044813 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.044852 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.044863 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.044888 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.044901 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.108210 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.108267 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.108210 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:01 crc kubenswrapper[4727]: E0929 10:24:01.108490 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:01 crc kubenswrapper[4727]: E0929 10:24:01.108670 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:01 crc kubenswrapper[4727]: E0929 10:24:01.109235 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.147470 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.147518 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.147533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.147562 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.147578 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.149898 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=82.149888744 podStartE2EDuration="1m22.149888744s" podCreationTimestamp="2025-09-29 10:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.149645267 +0000 UTC m=+111.322958649" watchObservedRunningTime="2025-09-29 10:24:01.149888744 +0000 UTC m=+111.323202106" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.164018 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.16400098 podStartE2EDuration="1m25.16400098s" podCreationTimestamp="2025-09-29 10:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.163757324 +0000 UTC m=+111.337070686" watchObservedRunningTime="2025-09-29 10:24:01.16400098 +0000 UTC m=+111.337314342" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.175367 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podStartSLOduration=84.175353217 podStartE2EDuration="1m24.175353217s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.174929067 +0000 UTC m=+111.348242439" watchObservedRunningTime="2025-09-29 10:24:01.175353217 +0000 UTC m=+111.348666579" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.226187 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6lmbz" podStartSLOduration=84.226173662 podStartE2EDuration="1m24.226173662s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.225889775 +0000 UTC m=+111.399203137" watchObservedRunningTime="2025-09-29 10:24:01.226173662 +0000 UTC m=+111.399487024" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.249969 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.250009 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.250022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.250035 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.250045 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.253850 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.253835771 podStartE2EDuration="1m24.253835771s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.251610925 +0000 UTC m=+111.424924287" watchObservedRunningTime="2025-09-29 10:24:01.253835771 +0000 UTC m=+111.427149133" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.285576 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mvx5d" podStartSLOduration=84.285556393 podStartE2EDuration="1m24.285556393s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.284595889 +0000 UTC m=+111.457909251" watchObservedRunningTime="2025-09-29 10:24:01.285556393 +0000 UTC m=+111.458869755" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.302015 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vg7wt" podStartSLOduration=84.301997349 podStartE2EDuration="1m24.301997349s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.300636484 +0000 UTC m=+111.473949846" watchObservedRunningTime="2025-09-29 10:24:01.301997349 +0000 UTC m=+111.475310711" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.330063 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.330029247 podStartE2EDuration="22.330029247s" podCreationTimestamp="2025-09-29 10:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.312243228 +0000 UTC m=+111.485556590" watchObservedRunningTime="2025-09-29 10:24:01.330029247 +0000 UTC m=+111.503342619" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.331042 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=61.331030842 podStartE2EDuration="1m1.331030842s" podCreationTimestamp="2025-09-29 10:23:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.329589716 +0000 UTC m=+111.502903088" watchObservedRunningTime="2025-09-29 10:24:01.331030842 +0000 UTC m=+111.504344214" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352828 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352857 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352865 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352879 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352889 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.352889 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-d8zdr" podStartSLOduration=83.352869174 podStartE2EDuration="1m23.352869174s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.352641219 +0000 UTC m=+111.525954621" watchObservedRunningTime="2025-09-29 10:24:01.352869174 +0000 UTC m=+111.526182536" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.455724 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.455783 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.455799 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.455823 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.455841 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.558414 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.558448 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.558458 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.558471 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.558480 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.660854 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.660911 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.660929 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.660951 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.660970 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.763586 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.763629 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.763645 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.763666 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.763680 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.866847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.866902 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.866921 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.866985 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.867250 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.969940 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.970005 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.970027 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.970072 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:01 crc kubenswrapper[4727]: I0929 10:24:01.970097 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:01Z","lastTransitionTime":"2025-09-29T10:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.073050 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.073101 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.073114 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.073131 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.073144 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.107849 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:02 crc kubenswrapper[4727]: E0929 10:24:02.107984 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.175832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.175863 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.175871 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.175885 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.175894 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.277897 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.277945 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.277957 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.277978 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.277991 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.380717 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.380763 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.380773 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.380789 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.380799 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.483150 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.483218 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.483235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.483261 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.483296 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.592916 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.592958 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.592967 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.592984 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.592996 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.696213 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.696260 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.696269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.696351 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.696362 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.799889 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.799939 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.799952 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.799973 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.799985 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.902309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.902391 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.902406 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.902424 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:02 crc kubenswrapper[4727]: I0929 10:24:02.902438 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:02Z","lastTransitionTime":"2025-09-29T10:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.005196 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.005235 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.005251 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.005273 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.005287 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.107533 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:03 crc kubenswrapper[4727]: E0929 10:24:03.107668 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.107927 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.107931 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:03 crc kubenswrapper[4727]: E0929 10:24:03.108062 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:03 crc kubenswrapper[4727]: E0929 10:24:03.108143 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.108882 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.109015 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.109116 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.109212 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.109299 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.214974 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.215014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.215024 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.215040 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.215050 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.317253 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.317563 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.317665 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.317761 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.317854 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.419832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.419872 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.419884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.419900 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.419910 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.522802 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.522840 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.522849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.522862 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.522871 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.625273 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.625309 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.625321 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.625359 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.625371 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.727633 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.727677 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.727688 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.727706 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.727718 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.830573 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.830603 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.830614 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.830630 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.830641 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.933634 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.933685 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.933699 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.933720 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:03 crc kubenswrapper[4727]: I0929 10:24:03.933735 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:03Z","lastTransitionTime":"2025-09-29T10:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.035758 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.035823 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.035836 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.035855 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.035867 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.108202 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:04 crc kubenswrapper[4727]: E0929 10:24:04.108537 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.138419 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.138462 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.138471 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.138488 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.138498 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.240480 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.240510 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.240519 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.240533 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.240542 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.342914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.342964 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.342977 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.342994 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.343004 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.444646 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.444687 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.444698 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.444714 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.444722 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.546972 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.547012 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.547022 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.547041 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.547053 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.650436 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.650478 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.650493 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.650514 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.650528 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.752842 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.752891 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.752905 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.752926 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.752938 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.856265 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.856306 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.856315 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.856332 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.856364 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.959446 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.959489 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.959499 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.959516 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:04 crc kubenswrapper[4727]: I0929 10:24:04.959530 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:04Z","lastTransitionTime":"2025-09-29T10:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.062820 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.062874 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.062887 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.062914 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.062940 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.108272 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.108417 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.108625 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:05 crc kubenswrapper[4727]: E0929 10:24:05.108770 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:05 crc kubenswrapper[4727]: E0929 10:24:05.108886 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:05 crc kubenswrapper[4727]: E0929 10:24:05.108985 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.165788 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.165847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.165863 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.165884 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.165899 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.268777 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.268831 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.268849 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.268869 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.268881 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.370682 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.371466 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.371508 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.371530 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.371544 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.475147 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.475184 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.475196 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.475213 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.475225 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.578603 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.578640 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.578650 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.578663 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.578674 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.680680 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.680739 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.680755 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.680778 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.680794 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.784029 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.784070 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.784079 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.784095 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.784104 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.886922 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.886999 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.887014 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.887033 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.887044 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.989553 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.989594 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.989603 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.989617 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:05 crc kubenswrapper[4727]: I0929 10:24:05.989627 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:05Z","lastTransitionTime":"2025-09-29T10:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.091864 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.091911 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.091924 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.091941 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.091953 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.108228 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:06 crc kubenswrapper[4727]: E0929 10:24:06.108326 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.195302 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.195356 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.195368 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.195386 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.195397 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.298211 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.298259 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.298269 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.298285 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.298296 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.400776 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.400821 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.400832 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.400847 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.400858 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.503023 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.503067 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.503079 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.503096 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.503107 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.605584 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.605613 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.605620 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.605635 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.605644 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.708699 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.708756 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.708772 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.708793 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.708809 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.740146 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.740197 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.740205 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.740219 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.740229 4727 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T10:24:06Z","lastTransitionTime":"2025-09-29T10:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.779257 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xwdcz" podStartSLOduration=89.779232776 podStartE2EDuration="1m29.779232776s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:01.397846551 +0000 UTC m=+111.571159913" watchObservedRunningTime="2025-09-29 10:24:06.779232776 +0000 UTC m=+116.952546148" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.780029 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf"] Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.780527 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.782885 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.782955 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.783008 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.783007 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.842665 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a789776-e91d-4398-9992-bd6b7b6c3309-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.842731 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.842774 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a789776-e91d-4398-9992-bd6b7b6c3309-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.842863 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a789776-e91d-4398-9992-bd6b7b6c3309-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.842917 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943866 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943915 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a789776-e91d-4398-9992-bd6b7b6c3309-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943940 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a789776-e91d-4398-9992-bd6b7b6c3309-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943965 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943984 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.943998 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a789776-e91d-4398-9992-bd6b7b6c3309-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.944060 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a789776-e91d-4398-9992-bd6b7b6c3309-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.944865 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a789776-e91d-4398-9992-bd6b7b6c3309-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.956858 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a789776-e91d-4398-9992-bd6b7b6c3309-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:06 crc kubenswrapper[4727]: I0929 10:24:06.959454 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a789776-e91d-4398-9992-bd6b7b6c3309-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5d5bf\" (UID: \"6a789776-e91d-4398-9992-bd6b7b6c3309\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.097321 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.108268 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.108315 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:07 crc kubenswrapper[4727]: E0929 10:24:07.108404 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.108294 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:07 crc kubenswrapper[4727]: E0929 10:24:07.108480 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:07 crc kubenswrapper[4727]: E0929 10:24:07.108526 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.797090 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" event={"ID":"6a789776-e91d-4398-9992-bd6b7b6c3309","Type":"ContainerStarted","Data":"67c18d2f291903e6c7250f12acf1bb3ba71ca4d611defe4f4bf570647479b8a0"} Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.797496 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" event={"ID":"6a789776-e91d-4398-9992-bd6b7b6c3309","Type":"ContainerStarted","Data":"38bc81161cd94ee59a95f48e30202ec9041dcab9723762874bbc6defcddff58b"} Sep 29 10:24:07 crc kubenswrapper[4727]: I0929 10:24:07.813143 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5d5bf" podStartSLOduration=90.813125559 podStartE2EDuration="1m30.813125559s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:07.813013496 +0000 UTC m=+117.986326868" watchObservedRunningTime="2025-09-29 10:24:07.813125559 +0000 UTC m=+117.986438931" Sep 29 10:24:08 crc kubenswrapper[4727]: I0929 10:24:08.107723 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:08 crc kubenswrapper[4727]: E0929 10:24:08.107871 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:09 crc kubenswrapper[4727]: I0929 10:24:09.108317 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:09 crc kubenswrapper[4727]: I0929 10:24:09.108480 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:09 crc kubenswrapper[4727]: I0929 10:24:09.108636 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:09 crc kubenswrapper[4727]: E0929 10:24:09.108723 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:09 crc kubenswrapper[4727]: E0929 10:24:09.109090 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:09 crc kubenswrapper[4727]: E0929 10:24:09.109190 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:10 crc kubenswrapper[4727]: I0929 10:24:10.108240 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:10 crc kubenswrapper[4727]: E0929 10:24:10.108703 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:10 crc kubenswrapper[4727]: I0929 10:24:10.110559 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:24:10 crc kubenswrapper[4727]: E0929 10:24:10.111773 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:24:11 crc kubenswrapper[4727]: E0929 10:24:11.052086 4727 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 29 10:24:11 crc kubenswrapper[4727]: I0929 10:24:11.107847 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:11 crc kubenswrapper[4727]: I0929 10:24:11.107898 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:11 crc kubenswrapper[4727]: I0929 10:24:11.107931 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:11 crc kubenswrapper[4727]: E0929 10:24:11.108875 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:11 crc kubenswrapper[4727]: E0929 10:24:11.109028 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:11 crc kubenswrapper[4727]: E0929 10:24:11.109206 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:11 crc kubenswrapper[4727]: E0929 10:24:11.213787 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:12 crc kubenswrapper[4727]: I0929 10:24:12.107612 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:12 crc kubenswrapper[4727]: E0929 10:24:12.107768 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:13 crc kubenswrapper[4727]: I0929 10:24:13.107602 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:13 crc kubenswrapper[4727]: I0929 10:24:13.107661 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:13 crc kubenswrapper[4727]: I0929 10:24:13.107599 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:13 crc kubenswrapper[4727]: E0929 10:24:13.107751 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:13 crc kubenswrapper[4727]: E0929 10:24:13.107814 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:13 crc kubenswrapper[4727]: E0929 10:24:13.107891 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:14 crc kubenswrapper[4727]: I0929 10:24:14.107881 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:14 crc kubenswrapper[4727]: E0929 10:24:14.108015 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:15 crc kubenswrapper[4727]: I0929 10:24:15.108237 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:15 crc kubenswrapper[4727]: I0929 10:24:15.108296 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:15 crc kubenswrapper[4727]: I0929 10:24:15.108559 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:15 crc kubenswrapper[4727]: E0929 10:24:15.108552 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:15 crc kubenswrapper[4727]: E0929 10:24:15.108660 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:15 crc kubenswrapper[4727]: E0929 10:24:15.108830 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:16 crc kubenswrapper[4727]: I0929 10:24:16.107893 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:16 crc kubenswrapper[4727]: E0929 10:24:16.108160 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:16 crc kubenswrapper[4727]: E0929 10:24:16.215496 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:17 crc kubenswrapper[4727]: I0929 10:24:17.107614 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:17 crc kubenswrapper[4727]: I0929 10:24:17.107667 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:17 crc kubenswrapper[4727]: I0929 10:24:17.107693 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:17 crc kubenswrapper[4727]: E0929 10:24:17.107837 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:17 crc kubenswrapper[4727]: E0929 10:24:17.107881 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:17 crc kubenswrapper[4727]: E0929 10:24:17.107945 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.107520 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:18 crc kubenswrapper[4727]: E0929 10:24:18.107873 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.833847 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/1.log" Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.834526 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/0.log" Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.834609 4727 generic.go:334] "Generic (PLEG): container finished" podID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" containerID="e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f" exitCode=1 Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.834659 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerDied","Data":"e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f"} Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.834707 4727 scope.go:117] "RemoveContainer" containerID="586c99a189912509131837aea7d60d92b1bc2cffc6aa7a37c97a19a40576f259" Sep 29 10:24:18 crc kubenswrapper[4727]: I0929 10:24:18.835501 4727 scope.go:117] "RemoveContainer" containerID="e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f" Sep 29 10:24:18 crc kubenswrapper[4727]: E0929 10:24:18.835927 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xwdcz_openshift-multus(9e3b98d9-a9fc-4bf3-8053-b7701c047d99)\"" pod="openshift-multus/multus-xwdcz" podUID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" Sep 29 10:24:19 crc kubenswrapper[4727]: I0929 10:24:19.108063 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:19 crc kubenswrapper[4727]: E0929 10:24:19.108489 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:19 crc kubenswrapper[4727]: I0929 10:24:19.108241 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:19 crc kubenswrapper[4727]: E0929 10:24:19.108600 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:19 crc kubenswrapper[4727]: I0929 10:24:19.108158 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:19 crc kubenswrapper[4727]: E0929 10:24:19.109269 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:19 crc kubenswrapper[4727]: I0929 10:24:19.839909 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/1.log" Sep 29 10:24:20 crc kubenswrapper[4727]: I0929 10:24:20.108028 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:20 crc kubenswrapper[4727]: E0929 10:24:20.108191 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:21 crc kubenswrapper[4727]: I0929 10:24:21.108568 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:21 crc kubenswrapper[4727]: E0929 10:24:21.110535 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:21 crc kubenswrapper[4727]: I0929 10:24:21.110601 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:21 crc kubenswrapper[4727]: E0929 10:24:21.111274 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:21 crc kubenswrapper[4727]: I0929 10:24:21.111494 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:21 crc kubenswrapper[4727]: E0929 10:24:21.111703 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:21 crc kubenswrapper[4727]: E0929 10:24:21.216319 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:22 crc kubenswrapper[4727]: I0929 10:24:22.107355 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:22 crc kubenswrapper[4727]: E0929 10:24:22.107542 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:22 crc kubenswrapper[4727]: I0929 10:24:22.108436 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:24:22 crc kubenswrapper[4727]: E0929 10:24:22.108675 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lqql2_openshift-ovn-kubernetes(cc578ace-78c8-4d17-a556-c6d6ceb149a4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" Sep 29 10:24:23 crc kubenswrapper[4727]: I0929 10:24:23.107820 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:23 crc kubenswrapper[4727]: I0929 10:24:23.107893 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:23 crc kubenswrapper[4727]: I0929 10:24:23.107854 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:23 crc kubenswrapper[4727]: E0929 10:24:23.108011 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:23 crc kubenswrapper[4727]: E0929 10:24:23.108206 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:23 crc kubenswrapper[4727]: E0929 10:24:23.108438 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:24 crc kubenswrapper[4727]: I0929 10:24:24.108130 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:24 crc kubenswrapper[4727]: E0929 10:24:24.108330 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:25 crc kubenswrapper[4727]: I0929 10:24:25.107749 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:25 crc kubenswrapper[4727]: I0929 10:24:25.107835 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:25 crc kubenswrapper[4727]: E0929 10:24:25.107901 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:25 crc kubenswrapper[4727]: E0929 10:24:25.108011 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:25 crc kubenswrapper[4727]: I0929 10:24:25.108099 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:25 crc kubenswrapper[4727]: E0929 10:24:25.108197 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:26 crc kubenswrapper[4727]: I0929 10:24:26.107932 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:26 crc kubenswrapper[4727]: E0929 10:24:26.108068 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:26 crc kubenswrapper[4727]: E0929 10:24:26.217639 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:27 crc kubenswrapper[4727]: I0929 10:24:27.108074 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:27 crc kubenswrapper[4727]: I0929 10:24:27.108094 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:27 crc kubenswrapper[4727]: E0929 10:24:27.108219 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:27 crc kubenswrapper[4727]: I0929 10:24:27.108235 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:27 crc kubenswrapper[4727]: E0929 10:24:27.108305 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:27 crc kubenswrapper[4727]: E0929 10:24:27.108404 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:28 crc kubenswrapper[4727]: I0929 10:24:28.107774 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:28 crc kubenswrapper[4727]: E0929 10:24:28.107893 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:29 crc kubenswrapper[4727]: I0929 10:24:29.108083 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:29 crc kubenswrapper[4727]: E0929 10:24:29.108215 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:29 crc kubenswrapper[4727]: I0929 10:24:29.108319 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:29 crc kubenswrapper[4727]: I0929 10:24:29.108484 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:29 crc kubenswrapper[4727]: E0929 10:24:29.108613 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:29 crc kubenswrapper[4727]: E0929 10:24:29.108818 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:30 crc kubenswrapper[4727]: I0929 10:24:30.107467 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:30 crc kubenswrapper[4727]: E0929 10:24:30.107617 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.107988 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:31 crc kubenswrapper[4727]: E0929 10:24:31.109445 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.109558 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.110047 4727 scope.go:117] "RemoveContainer" containerID="e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f" Sep 29 10:24:31 crc kubenswrapper[4727]: E0929 10:24:31.110033 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.110479 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:31 crc kubenswrapper[4727]: E0929 10:24:31.110607 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:31 crc kubenswrapper[4727]: E0929 10:24:31.218380 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.890181 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/1.log" Sep 29 10:24:31 crc kubenswrapper[4727]: I0929 10:24:31.890812 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerStarted","Data":"0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa"} Sep 29 10:24:32 crc kubenswrapper[4727]: I0929 10:24:32.107803 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:32 crc kubenswrapper[4727]: E0929 10:24:32.107975 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:33 crc kubenswrapper[4727]: I0929 10:24:33.108214 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:33 crc kubenswrapper[4727]: I0929 10:24:33.108368 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:33 crc kubenswrapper[4727]: E0929 10:24:33.108446 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:33 crc kubenswrapper[4727]: I0929 10:24:33.108396 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:33 crc kubenswrapper[4727]: E0929 10:24:33.108552 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:33 crc kubenswrapper[4727]: E0929 10:24:33.108695 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:34 crc kubenswrapper[4727]: I0929 10:24:34.107550 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:34 crc kubenswrapper[4727]: E0929 10:24:34.107779 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:35 crc kubenswrapper[4727]: I0929 10:24:35.107617 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:35 crc kubenswrapper[4727]: I0929 10:24:35.107666 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:35 crc kubenswrapper[4727]: E0929 10:24:35.107842 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:35 crc kubenswrapper[4727]: I0929 10:24:35.107876 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:35 crc kubenswrapper[4727]: E0929 10:24:35.108040 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:35 crc kubenswrapper[4727]: E0929 10:24:35.108185 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:36 crc kubenswrapper[4727]: I0929 10:24:36.107638 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:36 crc kubenswrapper[4727]: E0929 10:24:36.108069 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:36 crc kubenswrapper[4727]: I0929 10:24:36.108305 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:24:36 crc kubenswrapper[4727]: E0929 10:24:36.219189 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:37 crc kubenswrapper[4727]: I0929 10:24:37.108503 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:37 crc kubenswrapper[4727]: I0929 10:24:37.108636 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:37 crc kubenswrapper[4727]: I0929 10:24:37.108531 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:37 crc kubenswrapper[4727]: E0929 10:24:37.108777 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:37 crc kubenswrapper[4727]: E0929 10:24:37.109031 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:37 crc kubenswrapper[4727]: E0929 10:24:37.109281 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:38 crc kubenswrapper[4727]: I0929 10:24:38.107671 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:38 crc kubenswrapper[4727]: E0929 10:24:38.108025 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:38 crc kubenswrapper[4727]: I0929 10:24:38.922686 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/3.log" Sep 29 10:24:38 crc kubenswrapper[4727]: I0929 10:24:38.927190 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerStarted","Data":"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce"} Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.047795 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gttbj"] Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.048065 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:39 crc kubenswrapper[4727]: E0929 10:24:39.048317 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.107812 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.107894 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:39 crc kubenswrapper[4727]: E0929 10:24:39.107952 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.108047 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:39 crc kubenswrapper[4727]: E0929 10:24:39.108118 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:39 crc kubenswrapper[4727]: E0929 10:24:39.108414 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.930939 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:24:39 crc kubenswrapper[4727]: I0929 10:24:39.982172 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podStartSLOduration=122.982151682 podStartE2EDuration="2m2.982151682s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:39.97738135 +0000 UTC m=+150.150694752" watchObservedRunningTime="2025-09-29 10:24:39.982151682 +0000 UTC m=+150.155465054" Sep 29 10:24:41 crc kubenswrapper[4727]: I0929 10:24:41.107961 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:41 crc kubenswrapper[4727]: I0929 10:24:41.108021 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:41 crc kubenswrapper[4727]: E0929 10:24:41.109094 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:41 crc kubenswrapper[4727]: I0929 10:24:41.109158 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:41 crc kubenswrapper[4727]: I0929 10:24:41.109222 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:41 crc kubenswrapper[4727]: E0929 10:24:41.109428 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:41 crc kubenswrapper[4727]: E0929 10:24:41.109832 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:41 crc kubenswrapper[4727]: E0929 10:24:41.110227 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:41 crc kubenswrapper[4727]: E0929 10:24:41.221382 4727 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 10:24:43 crc kubenswrapper[4727]: I0929 10:24:43.107330 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:43 crc kubenswrapper[4727]: I0929 10:24:43.107439 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:43 crc kubenswrapper[4727]: E0929 10:24:43.107490 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:43 crc kubenswrapper[4727]: I0929 10:24:43.107520 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:43 crc kubenswrapper[4727]: I0929 10:24:43.107651 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:43 crc kubenswrapper[4727]: E0929 10:24:43.107696 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:43 crc kubenswrapper[4727]: E0929 10:24:43.107812 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:43 crc kubenswrapper[4727]: E0929 10:24:43.107929 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:44 crc kubenswrapper[4727]: I0929 10:24:44.983480 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:44 crc kubenswrapper[4727]: E0929 10:24:44.983628 4727 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:24:44 crc kubenswrapper[4727]: E0929 10:24:44.983713 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:26:46.983689721 +0000 UTC m=+277.157003123 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.084462 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.084644 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.084699 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.084742 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.084795 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:26:47.084757574 +0000 UTC m=+277.258070976 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.084916 4727 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.084931 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.084985 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085000 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085042 4727 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085064 4727 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085011 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 10:26:47.084986269 +0000 UTC m=+277.258299661 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085009 4727 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085175 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 10:26:47.085145193 +0000 UTC m=+277.258458585 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.085206 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 10:26:47.085190304 +0000 UTC m=+277.258503706 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.107635 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.107688 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.107847 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.107916 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.108119 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 10:24:45 crc kubenswrapper[4727]: I0929 10:24:45.108159 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.108256 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gttbj" podUID="59055eb3-6cad-4335-b100-39955f1c0500" Sep 29 10:24:45 crc kubenswrapper[4727]: E0929 10:24:45.108438 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.108313 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.108401 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.108414 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.108507 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.111417 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.111894 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.113653 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.114202 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.114242 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.115970 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.124106 4727 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.166769 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hr9hg"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.167730 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.178128 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.178922 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.179146 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.186637 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.186879 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.187039 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.187264 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.187432 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.188117 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.190118 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.190467 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.190628 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.191199 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.193440 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.193464 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lw586"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.193555 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.193816 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.194240 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.194830 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.194905 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.195234 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.195651 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.199224 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.200806 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.204308 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.205356 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.205642 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.205784 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.205942 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206127 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206329 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206545 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206647 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206888 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206946 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206553 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.207123 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.206887 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.207295 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.208366 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.210041 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.210606 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.211003 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.212455 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.212757 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.212864 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.213036 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.213083 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.213200 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.213291 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.213775 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-h9nz2"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.214259 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.214412 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.215026 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-q6kbd"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.215558 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.215875 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.215931 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.216562 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.216574 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.217396 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.234626 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-k7cws"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.235548 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.237777 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.237912 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.238883 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.239076 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.239583 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.239825 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.240007 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.240476 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.240588 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.240978 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.243467 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.243680 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.243854 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.243958 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.244061 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.244057 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.244662 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.260808 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.260938 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.261544 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.260955 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.262125 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.262198 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.262284 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.262640 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.262820 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.263545 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2qq2r"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.263955 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.264379 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.264618 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.264619 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.265158 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.267911 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vm4v4"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.268378 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lw586"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.268473 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.268890 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.269095 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.269840 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.271452 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.271766 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.272672 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.274666 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q4qs6"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.275322 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.275799 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-krfsh"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.275832 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.276234 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.276281 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.276974 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.277904 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.279202 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.279566 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280094 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280370 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280400 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280506 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280590 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280629 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280715 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280755 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280766 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280846 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280886 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280909 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280947 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280718 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281011 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280547 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281066 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281136 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281145 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281148 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.280849 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281289 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281319 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.281401 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.283701 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.283803 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.283901 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.283999 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.284362 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.284831 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.285211 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.285283 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.285481 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.295738 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.296755 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.297649 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.299148 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.308496 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.309552 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.311860 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.321565 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.322270 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.323218 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.323527 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-t4lhc"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324084 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324241 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94497a6a-de54-46ae-a609-12872ceb5b24-serving-cert\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324436 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfzt8\" (UniqueName: \"kubernetes.io/projected/6404327b-095b-435a-b953-b1c3b7a6f332-kube-api-access-bfzt8\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324588 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr27v\" (UniqueName: \"kubernetes.io/projected/81d87cad-5194-4768-87a0-90ac03416af8-kube-api-access-jr27v\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324783 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab27910-938e-4922-b9b0-b2cf44306cd6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324888 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324999 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325114 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325211 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td9zg\" (UniqueName: \"kubernetes.io/projected/fab27910-938e-4922-b9b0-b2cf44306cd6-kube-api-access-td9zg\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325312 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-audit\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325440 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-service-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.324233 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325953 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.325992 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326203 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-config\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326318 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326489 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326591 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326710 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326805 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-auth-proxy-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.326893 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb7n6\" (UniqueName: \"kubernetes.io/projected/c535a2d5-d067-4fd7-84d7-419e456bebfa-kube-api-access-gb7n6\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327003 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-serving-cert\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327074 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327490 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d87cad-5194-4768-87a0-90ac03416af8-serving-cert\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327625 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-client\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327741 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327909 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpnq6\" (UniqueName: \"kubernetes.io/projected/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-kube-api-access-dpnq6\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.327987 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328070 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-encryption-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328319 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpx8z\" (UniqueName: \"kubernetes.io/projected/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-kube-api-access-cpx8z\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328442 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zqvs\" (UniqueName: \"kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328574 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-client\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328713 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328818 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.328920 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329042 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab27910-938e-4922-b9b0-b2cf44306cd6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329218 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-trusted-ca\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329323 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329453 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329553 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-audit-dir\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329503 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329506 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.329651 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330109 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-dir\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330208 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-config\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330408 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330527 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj95w\" (UniqueName: \"kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330626 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330717 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-image-import-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330819 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-policies\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330921 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331029 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331120 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtpm4\" (UniqueName: \"kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331285 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331406 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330304 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331413 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.331997 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332107 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332261 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-serving-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332400 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332532 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-images\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332625 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-node-pullsecrets\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332722 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332818 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdxkx\" (UniqueName: \"kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.332935 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-serving-cert\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333078 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.330826 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333298 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6404327b-095b-435a-b953-b1c3b7a6f332-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333452 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-config\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333550 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c535a2d5-d067-4fd7-84d7-419e456bebfa-machine-approver-tls\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333646 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333736 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333837 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333935 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334044 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334124 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rz2nt"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334171 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-encryption-config\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334227 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldmck\" (UniqueName: \"kubernetes.io/projected/94497a6a-de54-46ae-a609-12872ceb5b24-kube-api-access-ldmck\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334308 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj5xg\" (UniqueName: \"kubernetes.io/projected/04eb68eb-eab0-4868-96e6-38781e313d3a-kube-api-access-dj5xg\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334387 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334572 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.333186 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334716 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.334981 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.335077 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.335495 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.335581 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.336237 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.344725 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.345505 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.346298 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.348253 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.348764 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-h9nz2"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.350319 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hr9hg"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.350532 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.359440 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zg47h"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.360116 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.360842 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.361089 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.361091 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.362932 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.363498 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-r9ffv"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.364120 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.364559 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-q6kbd"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.366104 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.367650 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-k7cws"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.369290 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2qq2r"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.370221 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.370873 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q4qs6"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.372173 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.373314 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vm4v4"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.376382 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.377860 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.380330 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.382974 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.384814 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.386471 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.388974 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.389898 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.391968 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.393442 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.398320 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.400372 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.404398 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.406954 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.408902 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rz2nt"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.410562 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.410667 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.411956 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-r9ffv"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.413425 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9mtzk"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.414590 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.414700 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.415835 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-lljzg"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.416598 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.416898 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.417999 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.419104 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.420187 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-krfsh"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.421265 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.422410 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9mtzk"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.423447 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-lljzg"] Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.430472 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.434947 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-service-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435077 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-audit\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435245 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435386 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435521 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-config\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435631 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435749 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435866 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435961 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435751 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-service-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.435809 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-audit\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436120 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436247 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-auth-proxy-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436413 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb7n6\" (UniqueName: \"kubernetes.io/projected/c535a2d5-d067-4fd7-84d7-419e456bebfa-kube-api-access-gb7n6\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436518 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-serving-cert\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436596 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d87cad-5194-4768-87a0-90ac03416af8-serving-cert\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436670 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-client\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436822 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpnq6\" (UniqueName: \"kubernetes.io/projected/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-kube-api-access-dpnq6\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436891 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-encryption-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436971 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpx8z\" (UniqueName: \"kubernetes.io/projected/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-kube-api-access-cpx8z\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437048 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zqvs\" (UniqueName: \"kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436450 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-config\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437065 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-auth-proxy-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437124 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-client\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436974 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436167 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.436757 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437632 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437691 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437714 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437731 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab27910-938e-4922-b9b0-b2cf44306cd6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437750 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-dir\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437816 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-trusted-ca\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437834 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437852 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437873 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-audit-dir\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437891 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437909 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-config\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437926 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437949 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj95w\" (UniqueName: \"kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437967 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.437981 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-image-import-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438001 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-policies\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438017 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438034 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438054 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtpm4\" (UniqueName: \"kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438072 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438094 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438121 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438185 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438217 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-serving-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438241 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438262 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-images\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438279 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-node-pullsecrets\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438294 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438312 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdxkx\" (UniqueName: \"kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438353 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-serving-cert\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438373 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438408 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6404327b-095b-435a-b953-b1c3b7a6f332-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438448 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-config\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438469 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c535a2d5-d067-4fd7-84d7-419e456bebfa-machine-approver-tls\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438488 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438506 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438521 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438538 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438554 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438575 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-encryption-config\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438591 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldmck\" (UniqueName: \"kubernetes.io/projected/94497a6a-de54-46ae-a609-12872ceb5b24-kube-api-access-ldmck\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438607 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj5xg\" (UniqueName: \"kubernetes.io/projected/04eb68eb-eab0-4868-96e6-38781e313d3a-kube-api-access-dj5xg\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438625 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94497a6a-de54-46ae-a609-12872ceb5b24-serving-cert\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438643 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfzt8\" (UniqueName: \"kubernetes.io/projected/6404327b-095b-435a-b953-b1c3b7a6f332-kube-api-access-bfzt8\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438653 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-node-pullsecrets\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438661 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438689 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr27v\" (UniqueName: \"kubernetes.io/projected/81d87cad-5194-4768-87a0-90ac03416af8-kube-api-access-jr27v\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438705 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab27910-938e-4922-b9b0-b2cf44306cd6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438754 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438772 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td9zg\" (UniqueName: \"kubernetes.io/projected/fab27910-938e-4922-b9b0-b2cf44306cd6-kube-api-access-td9zg\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438912 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.439013 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.439720 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-policies\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440296 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-config\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440354 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440649 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c535a2d5-d067-4fd7-84d7-419e456bebfa-config\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440694 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-image-import-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440980 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.440982 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-config\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.441596 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.441834 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442044 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-serving-cert\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.438372 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04eb68eb-eab0-4868-96e6-38781e313d3a-audit-dir\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442237 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-serving-ca\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442393 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-audit-dir\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442389 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442845 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94497a6a-de54-46ae-a609-12872ceb5b24-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.442851 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443088 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-encryption-config\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443141 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443408 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81d87cad-5194-4768-87a0-90ac03416af8-trusted-ca\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443757 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443784 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6404327b-095b-435a-b953-b1c3b7a6f332-images\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.443766 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-etcd-client\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.444234 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab27910-938e-4922-b9b0-b2cf44306cd6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.444322 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.444743 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.444863 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d87cad-5194-4768-87a0-90ac03416af8-serving-cert\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.445228 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.445303 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.445422 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.445558 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.445669 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446021 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c535a2d5-d067-4fd7-84d7-419e456bebfa-machine-approver-tls\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446204 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446295 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446533 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446583 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446693 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94497a6a-de54-46ae-a609-12872ceb5b24-serving-cert\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.446637 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.447043 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-serving-cert\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.447158 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6404327b-095b-435a-b953-b1c3b7a6f332-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.447262 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab27910-938e-4922-b9b0-b2cf44306cd6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.447768 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/04eb68eb-eab0-4868-96e6-38781e313d3a-etcd-client\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.447938 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-encryption-config\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.448735 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.450839 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.451976 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.456657 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.471482 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.490692 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.510665 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.531064 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.551980 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.571448 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.590393 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.610931 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.631756 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.649053 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.649698 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.650350 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.672208 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.691632 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.711514 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.743249 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.749990 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.771302 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.791150 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.811109 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.830740 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.851819 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.871084 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.911087 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.932409 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.951498 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.970832 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 29 10:24:47 crc kubenswrapper[4727]: I0929 10:24:47.990966 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.010848 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.031668 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.051604 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.072858 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.092680 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.110901 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.131601 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.150487 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.171630 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.190773 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.211407 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.231733 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.250982 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.271714 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.292534 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.311724 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.329527 4727 request.go:700] Waited for 1.005821889s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-sa-dockercfg-5xfcg&limit=500&resourceVersion=0 Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.331597 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.351468 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.371084 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.390523 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.411630 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.432524 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.451311 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.471620 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.491258 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.510892 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.530186 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.551520 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.573131 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.598210 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.611013 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.630847 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.651647 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.671313 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.691776 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.712809 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.732771 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.752573 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.772063 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.791619 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.824319 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.831786 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.853101 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.871578 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.891142 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.912165 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.931958 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.951558 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.971603 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 29 10:24:48 crc kubenswrapper[4727]: I0929 10:24:48.991077 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.012765 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.032238 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.052111 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.072751 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.091488 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.111143 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.131966 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.151772 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.170790 4727 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.191562 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.210949 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.230818 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.247030 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.247107 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.251051 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.271413 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.308596 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb7n6\" (UniqueName: \"kubernetes.io/projected/c535a2d5-d067-4fd7-84d7-419e456bebfa-kube-api-access-gb7n6\") pod \"machine-approver-56656f9798-pp54r\" (UID: \"c535a2d5-d067-4fd7-84d7-419e456bebfa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.325750 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpx8z\" (UniqueName: \"kubernetes.io/projected/d6a94e2a-c1ed-43c6-91d9-495a0bb673ce-kube-api-access-cpx8z\") pod \"apiserver-7bbb656c7d-ht2kk\" (UID: \"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.329611 4727 request.go:700] Waited for 1.892038911s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.353011 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zqvs\" (UniqueName: \"kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs\") pod \"oauth-openshift-558db77b4-nxzg5\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.367143 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpnq6\" (UniqueName: \"kubernetes.io/projected/a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44-kube-api-access-dpnq6\") pod \"cluster-samples-operator-665b6dd947-mwh6n\" (UID: \"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.385033 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td9zg\" (UniqueName: \"kubernetes.io/projected/fab27910-938e-4922-b9b0-b2cf44306cd6-kube-api-access-td9zg\") pod \"openshift-apiserver-operator-796bbdcf4f-vgphv\" (UID: \"fab27910-938e-4922-b9b0-b2cf44306cd6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.404380 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtpm4\" (UniqueName: \"kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4\") pod \"console-f9d7485db-fkfsn\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.422849 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdxkx\" (UniqueName: \"kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx\") pod \"controller-manager-879f6c89f-r97m9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.426672 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.444580 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj95w\" (UniqueName: \"kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w\") pod \"route-controller-manager-6576b87f9c-fcr2d\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.453785 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.465135 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr27v\" (UniqueName: \"kubernetes.io/projected/81d87cad-5194-4768-87a0-90ac03416af8-kube-api-access-jr27v\") pod \"console-operator-58897d9998-q6kbd\" (UID: \"81d87cad-5194-4768-87a0-90ac03416af8\") " pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.475179 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.484311 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.488520 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.489984 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj5xg\" (UniqueName: \"kubernetes.io/projected/04eb68eb-eab0-4868-96e6-38781e313d3a-kube-api-access-dj5xg\") pod \"apiserver-76f77b778f-hr9hg\" (UID: \"04eb68eb-eab0-4868-96e6-38781e313d3a\") " pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.496460 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.516571 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.520317 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldmck\" (UniqueName: \"kubernetes.io/projected/94497a6a-de54-46ae-a609-12872ceb5b24-kube-api-access-ldmck\") pod \"authentication-operator-69f744f599-h9nz2\" (UID: \"94497a6a-de54-46ae-a609-12872ceb5b24\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:49 crc kubenswrapper[4727]: W0929 10:24:49.545523 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc535a2d5_d067_4fd7_84d7_419e456bebfa.slice/crio-fc15cea14e0511f580d311a215df6c21e8d4be2a7950d99f606c2a900f23a240 WatchSource:0}: Error finding container fc15cea14e0511f580d311a215df6c21e8d4be2a7950d99f606c2a900f23a240: Status 404 returned error can't find the container with id fc15cea14e0511f580d311a215df6c21e8d4be2a7950d99f606c2a900f23a240 Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.546066 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfzt8\" (UniqueName: \"kubernetes.io/projected/6404327b-095b-435a-b953-b1c3b7a6f332-kube-api-access-bfzt8\") pod \"machine-api-operator-5694c8668f-lw586\" (UID: \"6404327b-095b-435a-b953-b1c3b7a6f332\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565191 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565244 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cd88cb-71a1-4e96-9091-152472b884a1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565272 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565295 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-cabundle\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565320 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/46124f8a-3360-4dfe-b468-5addce47a614-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565464 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565504 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8gnx\" (UniqueName: \"kubernetes.io/projected/64c27281-4623-451c-84b9-48b20e3a3a7d-kube-api-access-l8gnx\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565546 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhnr8\" (UniqueName: \"kubernetes.io/projected/f65657cc-1eae-43a4-82e2-65379bbed510-kube-api-access-zhnr8\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565563 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a2ce6052-f1b1-4f55-98f7-7e509ae66846-tmpfs\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565579 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565595 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e7cad1-1c3e-452e-979c-3ff3787050d1-config\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565618 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99pqg\" (UniqueName: \"kubernetes.io/projected/25b90489-d077-4da1-ae20-eb826bb3b189-kube-api-access-99pqg\") pod \"migrator-59844c95c7-2j69r\" (UID: \"25b90489-d077-4da1-ae20-eb826bb3b189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565632 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53f5f641-e101-4fa8-941e-540282897459-proxy-tls\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565680 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mznp9\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-kube-api-access-mznp9\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565723 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565742 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79209de4-57ff-43d1-9f46-aae7501d21c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565789 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d8d661-e0b8-47ff-8733-08c97804c66f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.565810 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hvpp\" (UniqueName: \"kubernetes.io/projected/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-kube-api-access-5hvpp\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567632 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567743 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/422a3020-82d4-48e2-ae59-9ab06061c5ee-available-featuregates\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567806 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-images\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567828 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlrxn\" (UniqueName: \"kubernetes.io/projected/39634023-d2e6-4b7e-a258-1d763c0bdaad-kube-api-access-zlrxn\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567879 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-service-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567944 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/274d4012-934c-426c-8902-20d19732cfea-service-ca-bundle\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567966 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.567987 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e7cad1-1c3e-452e-979c-3ff3787050d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.568236 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-webhook-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.568262 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gz2z\" (UniqueName: \"kubernetes.io/projected/274d4012-934c-426c-8902-20d19732cfea-kube-api-access-5gz2z\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.568333 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc42b\" (UniqueName: \"kubernetes.io/projected/23cd88cb-71a1-4e96-9091-152472b884a1-kube-api-access-cc42b\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.568384 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lll94\" (UniqueName: \"kubernetes.io/projected/46124f8a-3360-4dfe-b468-5addce47a614-kube-api-access-lll94\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.568647 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61d8d661-e0b8-47ff-8733-08c97804c66f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.569234 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.068739831 +0000 UTC m=+160.242053383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.569387 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.569443 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88be3001-6d99-49d8-b994-90178f3e64e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.569473 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/422a3020-82d4-48e2-ae59-9ab06061c5ee-serving-cert\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574452 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-metrics-tls\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574547 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-default-certificate\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574615 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574647 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-config\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574687 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-stats-auth\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574715 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574741 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebca6c99-76f7-4e93-9d6c-a91566decffc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574769 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbljq\" (UniqueName: \"kubernetes.io/projected/79209de4-57ff-43d1-9f46-aae7501d21c6-kube-api-access-tbljq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574811 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebca6c99-76f7-4e93-9d6c-a91566decffc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574886 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92zz8\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-kube-api-access-92zz8\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574955 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.574982 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-serving-cert\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575009 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-apiservice-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575036 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e7cad1-1c3e-452e-979c-3ff3787050d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575148 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575199 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39634023-d2e6-4b7e-a258-1d763c0bdaad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575222 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575263 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn46x\" (UniqueName: \"kubernetes.io/projected/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-kube-api-access-sn46x\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575286 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9sx7\" (UniqueName: \"kubernetes.io/projected/7f67f7b2-5a7d-421a-aa4a-333ef6d1e451-kube-api-access-h9sx7\") pod \"downloads-7954f5f757-2qq2r\" (UID: \"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451\") " pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575309 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j69br\" (UniqueName: \"kubernetes.io/projected/53f5f641-e101-4fa8-941e-540282897459-kube-api-access-j69br\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575397 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7fsk\" (UniqueName: \"kubernetes.io/projected/422a3020-82d4-48e2-ae59-9ab06061c5ee-kube-api-access-l7fsk\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575423 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-client\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575461 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-key\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575512 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88be3001-6d99-49d8-b994-90178f3e64e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575542 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlx8n\" (UniqueName: \"kubernetes.io/projected/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-kube-api-access-rlx8n\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575565 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph95m\" (UniqueName: \"kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575585 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-metrics-certs\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575630 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575653 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-srv-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575690 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575774 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575824 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23cd88cb-71a1-4e96-9091-152472b884a1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.575861 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsmdh\" (UniqueName: \"kubernetes.io/projected/a2ce6052-f1b1-4f55-98f7-7e509ae66846-kube-api-access-tsmdh\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.576144 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.576171 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88be3001-6d99-49d8-b994-90178f3e64e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.576252 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79209de4-57ff-43d1-9f46-aae7501d21c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.576277 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d8d661-e0b8-47ff-8733-08c97804c66f-config\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.576918 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrk6v\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.633681 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.674707 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.677984 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678253 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678520 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-node-bootstrap-token\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.678538 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.178518078 +0000 UTC m=+160.351831440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678557 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhnr8\" (UniqueName: \"kubernetes.io/projected/f65657cc-1eae-43a4-82e2-65379bbed510-kube-api-access-zhnr8\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678580 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a2ce6052-f1b1-4f55-98f7-7e509ae66846-tmpfs\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678608 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq5rt\" (UniqueName: \"kubernetes.io/projected/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-kube-api-access-bq5rt\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678631 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678654 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e7cad1-1c3e-452e-979c-3ff3787050d1-config\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678680 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-srv-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678704 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99pqg\" (UniqueName: \"kubernetes.io/projected/25b90489-d077-4da1-ae20-eb826bb3b189-kube-api-access-99pqg\") pod \"migrator-59844c95c7-2j69r\" (UID: \"25b90489-d077-4da1-ae20-eb826bb3b189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678727 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53f5f641-e101-4fa8-941e-540282897459-proxy-tls\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678750 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dgbf\" (UniqueName: \"kubernetes.io/projected/9e431cd8-2196-4aed-a4f6-0505374c7561-kube-api-access-8dgbf\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678776 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mznp9\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-kube-api-access-mznp9\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678798 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-profile-collector-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678828 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678851 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79209de4-57ff-43d1-9f46-aae7501d21c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678878 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d8d661-e0b8-47ff-8733-08c97804c66f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678902 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-mountpoint-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678926 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hvpp\" (UniqueName: \"kubernetes.io/projected/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-kube-api-access-5hvpp\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678954 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.678977 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/422a3020-82d4-48e2-ae59-9ab06061c5ee-available-featuregates\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679000 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-images\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679023 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlrxn\" (UniqueName: \"kubernetes.io/projected/39634023-d2e6-4b7e-a258-1d763c0bdaad-kube-api-access-zlrxn\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679046 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-service-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679090 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/274d4012-934c-426c-8902-20d19732cfea-service-ca-bundle\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679115 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679137 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e7cad1-1c3e-452e-979c-3ff3787050d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679160 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-webhook-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679185 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gz2z\" (UniqueName: \"kubernetes.io/projected/274d4012-934c-426c-8902-20d19732cfea-kube-api-access-5gz2z\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679212 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc42b\" (UniqueName: \"kubernetes.io/projected/23cd88cb-71a1-4e96-9091-152472b884a1-kube-api-access-cc42b\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679230 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lll94\" (UniqueName: \"kubernetes.io/projected/46124f8a-3360-4dfe-b468-5addce47a614-kube-api-access-lll94\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679248 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61d8d661-e0b8-47ff-8733-08c97804c66f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679266 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679285 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88be3001-6d99-49d8-b994-90178f3e64e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679304 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-csi-data-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679322 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/422a3020-82d4-48e2-ae59-9ab06061c5ee-serving-cert\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679363 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-certs\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679385 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8x2k\" (UniqueName: \"kubernetes.io/projected/8cca1787-b4dd-4b49-a610-3effd6117c44-kube-api-access-v8x2k\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679406 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-metrics-tls\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679421 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-default-certificate\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679455 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679461 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a2ce6052-f1b1-4f55-98f7-7e509ae66846-tmpfs\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679480 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-config\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679498 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-stats-auth\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679502 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.679526 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680250 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebca6c99-76f7-4e93-9d6c-a91566decffc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680282 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbljq\" (UniqueName: \"kubernetes.io/projected/79209de4-57ff-43d1-9f46-aae7501d21c6-kube-api-access-tbljq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680303 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e7cad1-1c3e-452e-979c-3ff3787050d1-config\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680320 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680396 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebca6c99-76f7-4e93-9d6c-a91566decffc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680422 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ae559a-06eb-4580-b4a8-c453e4e193a7-config\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680449 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92zz8\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-kube-api-access-92zz8\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680472 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680489 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-serving-cert\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680506 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-apiservice-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680530 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e7cad1-1c3e-452e-979c-3ff3787050d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680551 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680568 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-registration-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680587 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39634023-d2e6-4b7e-a258-1d763c0bdaad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680604 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680621 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn46x\" (UniqueName: \"kubernetes.io/projected/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-kube-api-access-sn46x\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680639 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9sx7\" (UniqueName: \"kubernetes.io/projected/7f67f7b2-5a7d-421a-aa4a-333ef6d1e451-kube-api-access-h9sx7\") pod \"downloads-7954f5f757-2qq2r\" (UID: \"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451\") " pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680656 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j69br\" (UniqueName: \"kubernetes.io/projected/53f5f641-e101-4fa8-941e-540282897459-kube-api-access-j69br\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680676 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7fsk\" (UniqueName: \"kubernetes.io/projected/422a3020-82d4-48e2-ae59-9ab06061c5ee-kube-api-access-l7fsk\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680692 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-client\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680712 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e00fd3e3-7070-497e-8370-ba2d427981dd-proxy-tls\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680730 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-key\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680750 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88be3001-6d99-49d8-b994-90178f3e64e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680768 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67m96\" (UniqueName: \"kubernetes.io/projected/a8ae559a-06eb-4580-b4a8-c453e4e193a7-kube-api-access-67m96\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680785 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlx8n\" (UniqueName: \"kubernetes.io/projected/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-kube-api-access-rlx8n\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680801 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph95m\" (UniqueName: \"kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680821 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-socket-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680836 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-metrics-certs\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680854 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzz7t\" (UniqueName: \"kubernetes.io/projected/65081bdc-ce5d-4a32-8c55-91316d66a83a-kube-api-access-mzz7t\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680875 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680892 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-srv-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680911 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680932 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tn29\" (UniqueName: \"kubernetes.io/projected/e00fd3e3-7070-497e-8370-ba2d427981dd-kube-api-access-8tn29\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680967 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fp9n\" (UniqueName: \"kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.680994 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681034 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23cd88cb-71a1-4e96-9091-152472b884a1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681069 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsmdh\" (UniqueName: \"kubernetes.io/projected/a2ce6052-f1b1-4f55-98f7-7e509ae66846-kube-api-access-tsmdh\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681092 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cca1787-b4dd-4b49-a610-3effd6117c44-cert\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681129 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2cnw\" (UniqueName: \"kubernetes.io/projected/dc06dadc-ca42-476d-ac08-656b17e587e8-kube-api-access-j2cnw\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681158 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681180 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681203 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ae559a-06eb-4580-b4a8-c453e4e193a7-serving-cert\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681224 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88be3001-6d99-49d8-b994-90178f3e64e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681244 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-plugins-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681279 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc06dadc-ca42-476d-ac08-656b17e587e8-config-volume\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681303 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79209de4-57ff-43d1-9f46-aae7501d21c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681326 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d8d661-e0b8-47ff-8733-08c97804c66f-config\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681407 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrk6v\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681431 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681454 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cd88cb-71a1-4e96-9091-152472b884a1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681476 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dc06dadc-ca42-476d-ac08-656b17e587e8-metrics-tls\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681500 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681524 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-cabundle\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681545 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/46124f8a-3360-4dfe-b468-5addce47a614-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681569 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e00fd3e3-7070-497e-8370-ba2d427981dd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681593 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681601 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebca6c99-76f7-4e93-9d6c-a91566decffc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.681664 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8gnx\" (UniqueName: \"kubernetes.io/projected/64c27281-4623-451c-84b9-48b20e3a3a7d-kube-api-access-l8gnx\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.686202 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-client\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.686290 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-webhook-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.692771 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.695051 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.695538 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88be3001-6d99-49d8-b994-90178f3e64e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.696242 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.696948 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-key\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.698666 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.699755 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-config\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.700150 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.700817 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.701095 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d8d661-e0b8-47ff-8733-08c97804c66f-config\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.701879 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.201862224 +0000 UTC m=+160.375175586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.706462 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.707140 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88be3001-6d99-49d8-b994-90178f3e64e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.707355 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53f5f641-e101-4fa8-941e-540282897459-images\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.708538 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-signing-cabundle\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.709121 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/274d4012-934c-426c-8902-20d19732cfea-service-ca-bundle\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.709280 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cd88cb-71a1-4e96-9091-152472b884a1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.712637 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.713273 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-default-certificate\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.713747 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-metrics-certs\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.715014 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.717113 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/422a3020-82d4-48e2-ae59-9ab06061c5ee-available-featuregates\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.717374 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.719532 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64c27281-4623-451c-84b9-48b20e3a3a7d-etcd-service-ca\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.722284 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79209de4-57ff-43d1-9f46-aae7501d21c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.723719 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d8d661-e0b8-47ff-8733-08c97804c66f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.723932 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2ce6052-f1b1-4f55-98f7-7e509ae66846-apiservice-cert\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.724616 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79209de4-57ff-43d1-9f46-aae7501d21c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.724644 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/274d4012-934c-426c-8902-20d19732cfea-stats-auth\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.725067 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-srv-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.725094 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39634023-d2e6-4b7e-a258-1d763c0bdaad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.725569 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e7cad1-1c3e-452e-979c-3ff3787050d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.725654 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/422a3020-82d4-48e2-ae59-9ab06061c5ee-serving-cert\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.725663 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebca6c99-76f7-4e93-9d6c-a91566decffc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.726058 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53f5f641-e101-4fa8-941e-540282897459-proxy-tls\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.726549 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-metrics-tls\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.728716 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.728938 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64c27281-4623-451c-84b9-48b20e3a3a7d-serving-cert\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.729087 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.729808 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65657cc-1eae-43a4-82e2-65379bbed510-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.732549 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23cd88cb-71a1-4e96-9091-152472b884a1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.733168 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhnr8\" (UniqueName: \"kubernetes.io/projected/f65657cc-1eae-43a4-82e2-65379bbed510-kube-api-access-zhnr8\") pod \"olm-operator-6b444d44fb-vvk6t\" (UID: \"f65657cc-1eae-43a4-82e2-65379bbed510\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.733507 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.733906 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.734118 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/46124f8a-3360-4dfe-b468-5addce47a614-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.743173 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" Sep 29 10:24:49 crc kubenswrapper[4727]: W0929 10:24:49.746582 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod357e2cd6_6aad_411f_a684_704708b9d291.slice/crio-a8d6c088981d6fb07f316741eabf436a9cab021621ef6812b0f5a8c5d20aaabc WatchSource:0}: Error finding container a8d6c088981d6fb07f316741eabf436a9cab021621ef6812b0f5a8c5d20aaabc: Status 404 returned error can't find the container with id a8d6c088981d6fb07f316741eabf436a9cab021621ef6812b0f5a8c5d20aaabc Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.761396 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e7cad1-1c3e-452e-979c-3ff3787050d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thw88\" (UID: \"01e7cad1-1c3e-452e-979c-3ff3787050d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.762047 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.777997 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbljq\" (UniqueName: \"kubernetes.io/projected/79209de4-57ff-43d1-9f46-aae7501d21c6-kube-api-access-tbljq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nw5vx\" (UID: \"79209de4-57ff-43d1-9f46-aae7501d21c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.784856 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785121 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-registration-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785189 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e00fd3e3-7070-497e-8370-ba2d427981dd-proxy-tls\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785213 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67m96\" (UniqueName: \"kubernetes.io/projected/a8ae559a-06eb-4580-b4a8-c453e4e193a7-kube-api-access-67m96\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785264 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-socket-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785295 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzz7t\" (UniqueName: \"kubernetes.io/projected/65081bdc-ce5d-4a32-8c55-91316d66a83a-kube-api-access-mzz7t\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785319 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tn29\" (UniqueName: \"kubernetes.io/projected/e00fd3e3-7070-497e-8370-ba2d427981dd-kube-api-access-8tn29\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785369 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fp9n\" (UniqueName: \"kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785414 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cca1787-b4dd-4b49-a610-3effd6117c44-cert\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785439 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2cnw\" (UniqueName: \"kubernetes.io/projected/dc06dadc-ca42-476d-ac08-656b17e587e8-kube-api-access-j2cnw\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785460 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785812 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-registration-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.785883 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-socket-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.785909 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.285891309 +0000 UTC m=+160.459204671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786012 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ae559a-06eb-4580-b4a8-c453e4e193a7-serving-cert\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786075 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-plugins-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786157 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc06dadc-ca42-476d-ac08-656b17e587e8-config-volume\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786226 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dc06dadc-ca42-476d-ac08-656b17e587e8-metrics-tls\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786266 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e00fd3e3-7070-497e-8370-ba2d427981dd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.786516 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-plugins-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.788091 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc06dadc-ca42-476d-ac08-656b17e587e8-config-volume\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.789115 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.789366 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e00fd3e3-7070-497e-8370-ba2d427981dd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.791909 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-node-bootstrap-token\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.791968 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq5rt\" (UniqueName: \"kubernetes.io/projected/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-kube-api-access-bq5rt\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792000 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-srv-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792365 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e00fd3e3-7070-497e-8370-ba2d427981dd-proxy-tls\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792455 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cca1787-b4dd-4b49-a610-3effd6117c44-cert\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792499 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dc06dadc-ca42-476d-ac08-656b17e587e8-metrics-tls\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792698 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dgbf\" (UniqueName: \"kubernetes.io/projected/9e431cd8-2196-4aed-a4f6-0505374c7561-kube-api-access-8dgbf\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792756 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-profile-collector-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792808 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792857 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-mountpoint-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.792984 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-csi-data-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.793033 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-certs\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.793061 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8x2k\" (UniqueName: \"kubernetes.io/projected/8cca1787-b4dd-4b49-a610-3effd6117c44-kube-api-access-v8x2k\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.793145 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.793179 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ae559a-06eb-4580-b4a8-c453e4e193a7-config\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.794430 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-csi-data-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.794101 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ae559a-06eb-4580-b4a8-c453e4e193a7-config\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.795052 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.295031062 +0000 UTC m=+160.468344504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.795484 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-mountpoint-dir\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.796732 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-srv-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.802297 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.804138 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-certs\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.805138 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.805799 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8gnx\" (UniqueName: \"kubernetes.io/projected/64c27281-4623-451c-84b9-48b20e3a3a7d-kube-api-access-l8gnx\") pod \"etcd-operator-b45778765-vm4v4\" (UID: \"64c27281-4623-451c-84b9-48b20e3a3a7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.809310 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ae559a-06eb-4580-b4a8-c453e4e193a7-serving-cert\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.812974 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-q6kbd"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.814705 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/65081bdc-ce5d-4a32-8c55-91316d66a83a-node-bootstrap-token\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.822657 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.824004 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61d8d661-e0b8-47ff-8733-08c97804c66f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m4jkd\" (UID: \"61d8d661-e0b8-47ff-8733-08c97804c66f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.824218 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9e431cd8-2196-4aed-a4f6-0505374c7561-profile-collector-cert\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.836173 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gz2z\" (UniqueName: \"kubernetes.io/projected/274d4012-934c-426c-8902-20d19732cfea-kube-api-access-5gz2z\") pod \"router-default-5444994796-t4lhc\" (UID: \"274d4012-934c-426c-8902-20d19732cfea\") " pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.858690 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.859425 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc42b\" (UniqueName: \"kubernetes.io/projected/23cd88cb-71a1-4e96-9091-152472b884a1-kube-api-access-cc42b\") pod \"kube-storage-version-migrator-operator-b67b599dd-7ws9x\" (UID: \"23cd88cb-71a1-4e96-9091-152472b884a1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.873129 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lll94\" (UniqueName: \"kubernetes.io/projected/46124f8a-3360-4dfe-b468-5addce47a614-kube-api-access-lll94\") pod \"control-plane-machine-set-operator-78cbb6b69f-f7hn5\" (UID: \"46124f8a-3360-4dfe-b468-5addce47a614\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.881220 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.889222 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.893402 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlx8n\" (UniqueName: \"kubernetes.io/projected/194f9707-e39e-4a4a-bda7-c6c69b0b2e7c-kube-api-access-rlx8n\") pod \"multus-admission-controller-857f4d67dd-q4qs6\" (UID: \"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.893954 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:49 crc kubenswrapper[4727]: E0929 10:24:49.894747 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.394731933 +0000 UTC m=+160.568045296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.896811 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.914638 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph95m\" (UniqueName: \"kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m\") pod \"marketplace-operator-79b997595-65ccr\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.918781 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.925730 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.931808 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.951467 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.952700 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsmdh\" (UniqueName: \"kubernetes.io/projected/a2ce6052-f1b1-4f55-98f7-7e509ae66846-kube-api-access-tsmdh\") pod \"packageserver-d55dfcdfc-ql82q\" (UID: \"a2ce6052-f1b1-4f55-98f7-7e509ae66846\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.968453 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrk6v\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.972236 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" event={"ID":"f1fb18d6-876c-4a19-9aab-fa0cfaccd975","Type":"ContainerStarted","Data":"52f41e3b6e92d4ad2bd85e390c09497738b6196431f9c43d0b3821a4d9504dac"} Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.974183 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" event={"ID":"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce","Type":"ContainerStarted","Data":"786ea48feef22edace8e51398ada81be801f38407b3d862a702722e0031d31d7"} Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.981052 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.990124 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" event={"ID":"81d87cad-5194-4768-87a0-90ac03416af8","Type":"ContainerStarted","Data":"38393e9dbd25a084ae50371217e272ecf4bfe39aa861609c10b655919e3ca259"} Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.992580 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mznp9\" (UniqueName: \"kubernetes.io/projected/d6c27c2f-cae2-40bc-8889-9a14daedc8ed-kube-api-access-mznp9\") pod \"cluster-image-registry-operator-dc59b4c8b-l7ggf\" (UID: \"d6c27c2f-cae2-40bc-8889-9a14daedc8ed\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:49 crc kubenswrapper[4727]: I0929 10:24:49.992890 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.000103 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.000491 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.500469356 +0000 UTC m=+160.673782888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.003458 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" event={"ID":"c535a2d5-d067-4fd7-84d7-419e456bebfa","Type":"ContainerStarted","Data":"fc15cea14e0511f580d311a215df6c21e8d4be2a7950d99f606c2a900f23a240"} Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.007497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" event={"ID":"357e2cd6-6aad-411f-a684-704708b9d291","Type":"ContainerStarted","Data":"a8d6c088981d6fb07f316741eabf436a9cab021621ef6812b0f5a8c5d20aaabc"} Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.010902 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99pqg\" (UniqueName: \"kubernetes.io/projected/25b90489-d077-4da1-ae20-eb826bb3b189-kube-api-access-99pqg\") pod \"migrator-59844c95c7-2j69r\" (UID: \"25b90489-d077-4da1-ae20-eb826bb3b189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.011246 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.026432 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92zz8\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-kube-api-access-92zz8\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.049968 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j69br\" (UniqueName: \"kubernetes.io/projected/53f5f641-e101-4fa8-941e-540282897459-kube-api-access-j69br\") pod \"machine-config-operator-74547568cd-z2jgn\" (UID: \"53f5f641-e101-4fa8-941e-540282897459\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.069477 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebca6c99-76f7-4e93-9d6c-a91566decffc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zn6h\" (UID: \"ebca6c99-76f7-4e93-9d6c-a91566decffc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.088517 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88be3001-6d99-49d8-b994-90178f3e64e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jc8lh\" (UID: \"88be3001-6d99-49d8-b994-90178f3e64e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.101007 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.101425 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.601407156 +0000 UTC m=+160.774720518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.109593 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.127777 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7fsk\" (UniqueName: \"kubernetes.io/projected/422a3020-82d4-48e2-ae59-9ab06061c5ee-kube-api-access-l7fsk\") pod \"openshift-config-operator-7777fb866f-k7cws\" (UID: \"422a3020-82d4-48e2-ae59-9ab06061c5ee\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.132723 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.155296 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hvpp\" (UniqueName: \"kubernetes.io/projected/9fe1e500-bf28-4968-bca6-4e5e53ee2a1a-kube-api-access-5hvpp\") pod \"dns-operator-744455d44c-krfsh\" (UID: \"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.170742 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlrxn\" (UniqueName: \"kubernetes.io/projected/39634023-d2e6-4b7e-a258-1d763c0bdaad-kube-api-access-zlrxn\") pod \"package-server-manager-789f6589d5-fzb2k\" (UID: \"39634023-d2e6-4b7e-a258-1d763c0bdaad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.184064 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.188360 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9sx7\" (UniqueName: \"kubernetes.io/projected/7f67f7b2-5a7d-421a-aa4a-333ef6d1e451-kube-api-access-h9sx7\") pod \"downloads-7954f5f757-2qq2r\" (UID: \"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451\") " pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.202274 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.202788 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.203277 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.703261757 +0000 UTC m=+160.876575129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.207887 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn46x\" (UniqueName: \"kubernetes.io/projected/1030d2bc-8c0b-42b6-88da-99ca8f63a1ff-kube-api-access-sn46x\") pod \"service-ca-9c57cc56f-rz2nt\" (UID: \"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.211798 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.226328 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzz7t\" (UniqueName: \"kubernetes.io/projected/65081bdc-ce5d-4a32-8c55-91316d66a83a-kube-api-access-mzz7t\") pod \"machine-config-server-zg47h\" (UID: \"65081bdc-ce5d-4a32-8c55-91316d66a83a\") " pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.236736 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.245255 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.261214 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.261464 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.272941 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.276946 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.277006 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.277950 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q4qs6"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.278087 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tn29\" (UniqueName: \"kubernetes.io/projected/e00fd3e3-7070-497e-8370-ba2d427981dd-kube-api-access-8tn29\") pod \"machine-config-controller-84d6567774-zpkh2\" (UID: \"e00fd3e3-7070-497e-8370-ba2d427981dd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.290455 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fp9n\" (UniqueName: \"kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n\") pod \"collect-profiles-29319015-l24x8\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.300986 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.303754 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.304045 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.804019263 +0000 UTC m=+160.977332625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.305906 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2cnw\" (UniqueName: \"kubernetes.io/projected/dc06dadc-ca42-476d-ac08-656b17e587e8-kube-api-access-j2cnw\") pod \"dns-default-lljzg\" (UID: \"dc06dadc-ca42-476d-ac08-656b17e587e8\") " pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.317984 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.323247 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.326075 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq5rt\" (UniqueName: \"kubernetes.io/projected/7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee-kube-api-access-bq5rt\") pod \"csi-hostpathplugin-9mtzk\" (UID: \"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee\") " pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.334654 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67m96\" (UniqueName: \"kubernetes.io/projected/a8ae559a-06eb-4580-b4a8-c453e4e193a7-kube-api-access-67m96\") pod \"service-ca-operator-777779d784-vl9k7\" (UID: \"a8ae559a-06eb-4580-b4a8-c453e4e193a7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.355229 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zg47h" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.355238 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.359856 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vm4v4"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.361691 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lw586"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.362728 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hr9hg"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.363438 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8x2k\" (UniqueName: \"kubernetes.io/projected/8cca1787-b4dd-4b49-a610-3effd6117c44-kube-api-access-v8x2k\") pod \"ingress-canary-r9ffv\" (UID: \"8cca1787-b4dd-4b49-a610-3effd6117c44\") " pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.372845 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dgbf\" (UniqueName: \"kubernetes.io/projected/9e431cd8-2196-4aed-a4f6-0505374c7561-kube-api-access-8dgbf\") pod \"catalog-operator-68c6474976-x7jzw\" (UID: \"9e431cd8-2196-4aed-a4f6-0505374c7561\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:50 crc kubenswrapper[4727]: W0929 10:24:50.379882 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfab27910_938e_4922_b9b0_b2cf44306cd6.slice/crio-7df113d422306db82ae7ba5f7da7e83d9be9b5e289d2fb53928b0da9e64c3308 WatchSource:0}: Error finding container 7df113d422306db82ae7ba5f7da7e83d9be9b5e289d2fb53928b0da9e64c3308: Status 404 returned error can't find the container with id 7df113d422306db82ae7ba5f7da7e83d9be9b5e289d2fb53928b0da9e64c3308 Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.383993 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-h9nz2"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.386039 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" Sep 29 10:24:50 crc kubenswrapper[4727]: W0929 10:24:50.387677 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6404327b_095b_435a_b953_b1c3b7a6f332.slice/crio-b07a0468b3728613b34381cc85377e7ac9a6558c1f9572aed65fe326c43d5481 WatchSource:0}: Error finding container b07a0468b3728613b34381cc85377e7ac9a6558c1f9572aed65fe326c43d5481: Status 404 returned error can't find the container with id b07a0468b3728613b34381cc85377e7ac9a6558c1f9572aed65fe326c43d5481 Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.393884 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-lljzg" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.405970 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.406002 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.406362 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:50.906331585 +0000 UTC m=+161.079644947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.438233 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.507191 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.507401 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.007370578 +0000 UTC m=+161.180683970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.507850 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.508362 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.008351021 +0000 UTC m=+161.181664393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.575215 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.608591 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.608753 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.108718197 +0000 UTC m=+161.282031559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.608793 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.609200 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.109183888 +0000 UTC m=+161.282497240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.633025 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.640391 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.654696 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-r9ffv" Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.673068 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.700348 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.710292 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.710635 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.21059811 +0000 UTC m=+161.383911472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.710826 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.711244 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.211233625 +0000 UTC m=+161.384547157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.736590 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.771582 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.812422 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.812647 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.312618525 +0000 UTC m=+161.485931887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.812875 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.813235 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.313217259 +0000 UTC m=+161.486530691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:50 crc kubenswrapper[4727]: I0929 10:24:50.913659 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:50 crc kubenswrapper[4727]: E0929 10:24:50.914885 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.414865456 +0000 UTC m=+161.588178828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.011828 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" event={"ID":"64c27281-4623-451c-84b9-48b20e3a3a7d","Type":"ContainerStarted","Data":"5d5139810c506280f546146e4a2ec0e4def9bb75b54f0e3f3c3d2304c73ed25b"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.012619 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" event={"ID":"94497a6a-de54-46ae-a609-12872ceb5b24","Type":"ContainerStarted","Data":"789725e65be81627a5113db9c6279ae132ebc85ab5b9d1dcaf933efc00e0174e"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.013400 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" event={"ID":"305b863d-a41d-437c-986b-04a2bb52fde9","Type":"ContainerStarted","Data":"6466e6b61047c567a2a27904b29c879371bc07f697fd1f1bea4ef0d034aa5c77"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.014891 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" event={"ID":"23cd88cb-71a1-4e96-9091-152472b884a1","Type":"ContainerStarted","Data":"e07f7f7d0e2f7216727bafe4021a36bdd84f09b4899b20b9801f9008452eb864"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.015904 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.016189 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" event={"ID":"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44","Type":"ContainerStarted","Data":"5f88ed837e596d8d89a44bac7b9cf1b2b9f275e00fe9f4f31f399d368221972a"} Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.016384 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.516360969 +0000 UTC m=+161.689674371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.017476 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" event={"ID":"6404327b-095b-435a-b953-b1c3b7a6f332","Type":"ContainerStarted","Data":"b07a0468b3728613b34381cc85377e7ac9a6558c1f9572aed65fe326c43d5481"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.018452 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" event={"ID":"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c","Type":"ContainerStarted","Data":"97d9bceeba55d865210697ad2749466f5d899d7ed977836634c6e557ba9e357a"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.023633 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fkfsn" event={"ID":"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5","Type":"ContainerStarted","Data":"d9a0154e7f8d1c04ed7fceef7a1218fb4fec95c7923e37f198ced4b7291b061b"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.024475 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" event={"ID":"79209de4-57ff-43d1-9f46-aae7501d21c6","Type":"ContainerStarted","Data":"29f3e6459f1648fc32754a9a3d9191c71bab69ee30260f28f353fe17ff4d4334"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.025311 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" event={"ID":"04eb68eb-eab0-4868-96e6-38781e313d3a","Type":"ContainerStarted","Data":"d9ad3952c1c86598d0b7b12e1dc3f9fd7f149873822ffe1de101cff934915e3a"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.026249 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" event={"ID":"fab27910-938e-4922-b9b0-b2cf44306cd6","Type":"ContainerStarted","Data":"7df113d422306db82ae7ba5f7da7e83d9be9b5e289d2fb53928b0da9e64c3308"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.026948 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t4lhc" event={"ID":"274d4012-934c-426c-8902-20d19732cfea","Type":"ContainerStarted","Data":"f8af297ab47c7ad5504f15fe9ffcf2f3d82584eca5986abc419a690f72068439"} Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.042145 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.116878 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.117234 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.617196757 +0000 UTC m=+161.790510129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.117417 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.117933 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.617918524 +0000 UTC m=+161.791231906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.125563 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01e7cad1_1c3e_452e_979c_3ff3787050d1.slice/crio-49e8cd94d1cff4ad65d7bea6bcce577f56f5b0fdc7b21b40299202d0d2cee921 WatchSource:0}: Error finding container 49e8cd94d1cff4ad65d7bea6bcce577f56f5b0fdc7b21b40299202d0d2cee921: Status 404 returned error can't find the container with id 49e8cd94d1cff4ad65d7bea6bcce577f56f5b0fdc7b21b40299202d0d2cee921 Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.127679 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46124f8a_3360_4dfe_b468_5addce47a614.slice/crio-f7426adbf4a46fa9dac6d57732028a73abce392c7739d83538560bbe4d02efc0 WatchSource:0}: Error finding container f7426adbf4a46fa9dac6d57732028a73abce392c7739d83538560bbe4d02efc0: Status 404 returned error can't find the container with id f7426adbf4a46fa9dac6d57732028a73abce392c7739d83538560bbe4d02efc0 Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.129871 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61d8d661_e0b8_47ff_8733_08c97804c66f.slice/crio-47b1719d1bef62e6fd083daceba13c04121899c4cd75f230bedb955ee3c69965 WatchSource:0}: Error finding container 47b1719d1bef62e6fd083daceba13c04121899c4cd75f230bedb955ee3c69965: Status 404 returned error can't find the container with id 47b1719d1bef62e6fd083daceba13c04121899c4cd75f230bedb955ee3c69965 Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.133158 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73a9e11c_a13e_4167_9744_d7427e583d89.slice/crio-a4fd05c74d62fc64a31b2c813e9aae1011bbd26c7633a5a86e201b37e726bfc9 WatchSource:0}: Error finding container a4fd05c74d62fc64a31b2c813e9aae1011bbd26c7633a5a86e201b37e726bfc9: Status 404 returned error can't find the container with id a4fd05c74d62fc64a31b2c813e9aae1011bbd26c7633a5a86e201b37e726bfc9 Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.144145 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88be3001_6d99_49d8_b994_90178f3e64e0.slice/crio-cf1c526ec68dddfae998fe3c5317eed1caac10cf8b06ce0c0ef6cfc9e44b0469 WatchSource:0}: Error finding container cf1c526ec68dddfae998fe3c5317eed1caac10cf8b06ce0c0ef6cfc9e44b0469: Status 404 returned error can't find the container with id cf1c526ec68dddfae998fe3c5317eed1caac10cf8b06ce0c0ef6cfc9e44b0469 Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.220899 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.221555 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.721535027 +0000 UTC m=+161.894848389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.286148 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.322825 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.323395 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.823367548 +0000 UTC m=+161.996681080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.367164 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rz2nt"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.403089 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.404851 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-krfsh"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.424300 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.424447 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.924425291 +0000 UTC m=+162.097738663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.424751 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.425102 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:51.925092456 +0000 UTC m=+162.098405818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.525818 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.525962 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.025945154 +0000 UTC m=+162.199258516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.526229 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.526566 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.026558319 +0000 UTC m=+162.199871681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.627140 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.627591 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.127573781 +0000 UTC m=+162.300887143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.729114 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.729672 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.229659348 +0000 UTC m=+162.402972710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.755846 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-lljzg"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.798716 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.824873 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.830214 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.830394 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.330370393 +0000 UTC m=+162.503683755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.830578 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.830862 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.330850234 +0000 UTC m=+162.504163586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.874670 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9mtzk"] Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.931601 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.931785 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.431751743 +0000 UTC m=+162.605065125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: I0929 10:24:51.932290 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:51 crc kubenswrapper[4727]: E0929 10:24:51.932726 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.432716256 +0000 UTC m=+162.606029618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:51 crc kubenswrapper[4727]: W0929 10:24:51.950004 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6c27c2f_cae2_40bc_8889_9a14daedc8ed.slice/crio-778f357b2c6ab152d7fb21ca783ed4edccc8d93b21ed69d66a9f80c2f289e69b WatchSource:0}: Error finding container 778f357b2c6ab152d7fb21ca783ed4edccc8d93b21ed69d66a9f80c2f289e69b: Status 404 returned error can't find the container with id 778f357b2c6ab152d7fb21ca783ed4edccc8d93b21ed69d66a9f80c2f289e69b Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.033148 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.033284 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.533253707 +0000 UTC m=+162.706567069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.033822 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.034145 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.534130157 +0000 UTC m=+162.707443519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.034451 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" event={"ID":"f1fb18d6-876c-4a19-9aab-fa0cfaccd975","Type":"ContainerStarted","Data":"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.060612 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lljzg" event={"ID":"dc06dadc-ca42-476d-ac08-656b17e587e8","Type":"ContainerStarted","Data":"52b4df585f5538f9c12cdd72712a460f61c9f0d6929e8363c50d040be618fc40"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.086136 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" event={"ID":"73a9e11c-a13e-4167-9744-d7427e583d89","Type":"ContainerStarted","Data":"a4fd05c74d62fc64a31b2c813e9aae1011bbd26c7633a5a86e201b37e726bfc9"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.105190 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" event={"ID":"f65657cc-1eae-43a4-82e2-65379bbed510","Type":"ContainerStarted","Data":"b45727ba44d662c8ae47fc323f51407c6079916b06bdf96bac6f3c5c6a3e183f"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.119681 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" event={"ID":"d6c27c2f-cae2-40bc-8889-9a14daedc8ed","Type":"ContainerStarted","Data":"778f357b2c6ab152d7fb21ca783ed4edccc8d93b21ed69d66a9f80c2f289e69b"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.123180 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" event={"ID":"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2","Type":"ContainerStarted","Data":"6fd452c1006309f0a895c3f2f487e80cc876b4a7016433f0a0bab2aaa285500c"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.130017 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" event={"ID":"c535a2d5-d067-4fd7-84d7-419e456bebfa","Type":"ContainerStarted","Data":"24282971a076e4906ab7d5db3adf954740ced177e4ad5f5f4c25dad938bef873"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.131757 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zg47h" event={"ID":"65081bdc-ce5d-4a32-8c55-91316d66a83a","Type":"ContainerStarted","Data":"5aa5a3b6ecc7d73164ed491ff7423557af909fb0985c7de2c820826cc9b51598"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.133253 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" event={"ID":"46124f8a-3360-4dfe-b468-5addce47a614","Type":"ContainerStarted","Data":"f7426adbf4a46fa9dac6d57732028a73abce392c7739d83538560bbe4d02efc0"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.143491 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.144701 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.644655812 +0000 UTC m=+162.817969174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.149060 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" event={"ID":"61d8d661-e0b8-47ff-8733-08c97804c66f","Type":"ContainerStarted","Data":"47b1719d1bef62e6fd083daceba13c04121899c4cd75f230bedb955ee3c69965"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.162478 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" event={"ID":"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a","Type":"ContainerStarted","Data":"715af97a8ecd7deac0595daf166fe1c203d81639c94c3c8c55926f84efa7b21e"} Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.163490 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a94e2a_c1ed_43c6_91d9_495a0bb673ce.slice/crio-e20e085a3621c49d35088be720ff2ba82b944dd9f1f90eebe382a97fd8087dd1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a94e2a_c1ed_43c6_91d9_495a0bb673ce.slice/crio-conmon-e20e085a3621c49d35088be720ff2ba82b944dd9f1f90eebe382a97fd8087dd1.scope\": RecentStats: unable to find data in memory cache]" Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.167784 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" event={"ID":"01e7cad1-1c3e-452e-979c-3ff3787050d1","Type":"ContainerStarted","Data":"49e8cd94d1cff4ad65d7bea6bcce577f56f5b0fdc7b21b40299202d0d2cee921"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.175928 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" event={"ID":"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff","Type":"ContainerStarted","Data":"f567855d67a8769613038fd6a6032fb96d4e46d8d7d311556a4c9e2ea86e1775"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.183015 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" event={"ID":"357e2cd6-6aad-411f-a684-704708b9d291","Type":"ContainerStarted","Data":"ace6edc6233d2af7cd6f36b3fdcddc0fd67a146067a8963d3b0466a0e852d99a"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.221000 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" event={"ID":"88be3001-6d99-49d8-b994-90178f3e64e0","Type":"ContainerStarted","Data":"cf1c526ec68dddfae998fe3c5317eed1caac10cf8b06ce0c0ef6cfc9e44b0469"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.226976 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" event={"ID":"53f5f641-e101-4fa8-941e-540282897459","Type":"ContainerStarted","Data":"b01fde85ad2a7bdc296759f40dfc3adba2b0110bff66ba146f040df12804c24e"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.231508 4727 generic.go:334] "Generic (PLEG): container finished" podID="d6a94e2a-c1ed-43c6-91d9-495a0bb673ce" containerID="e20e085a3621c49d35088be720ff2ba82b944dd9f1f90eebe382a97fd8087dd1" exitCode=0 Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.231571 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" event={"ID":"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce","Type":"ContainerDied","Data":"e20e085a3621c49d35088be720ff2ba82b944dd9f1f90eebe382a97fd8087dd1"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.236640 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" event={"ID":"81d87cad-5194-4768-87a0-90ac03416af8","Type":"ContainerStarted","Data":"00f0fd6fd0c3995f2e4ac6bcd25d2b92f5322e70d02b300f8dfc6cbda7079097"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.238484 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" event={"ID":"ebca6c99-76f7-4e93-9d6c-a91566decffc","Type":"ContainerStarted","Data":"b522a88aa82b51e0a7ab5d82e0a0f94ea501f6a3056dab16d1fa495e3aaa044e"} Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.246317 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.246722 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.746705618 +0000 UTC m=+162.920018980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.299848 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.347536 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.347789 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.847755761 +0000 UTC m=+163.021069123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.348169 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.348604 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.8485696 +0000 UTC m=+163.021882962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.352527 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.355492 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.362859 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.381895 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.448919 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.449080 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.949059329 +0000 UTC m=+163.122372691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.449271 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.449681 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:52.949669524 +0000 UTC m=+163.122982886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.534921 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-r9ffv"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.538133 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2qq2r"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.540444 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.542305 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-k7cws"] Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.549879 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.550110 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.050076081 +0000 UTC m=+163.223389443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.550256 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.550585 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.050575283 +0000 UTC m=+163.223888645 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.640963 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ce6052_f1b1_4f55_98f7_7e509ae66846.slice/crio-9a4044e618f09a985b76fa0491d9c1af80fbf385a950267102c52265bc65021b WatchSource:0}: Error finding container 9a4044e618f09a985b76fa0491d9c1af80fbf385a950267102c52265bc65021b: Status 404 returned error can't find the container with id 9a4044e618f09a985b76fa0491d9c1af80fbf385a950267102c52265bc65021b Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.645239 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39634023_d2e6_4b7e_a258_1d763c0bdaad.slice/crio-66a3b53c0e30a7b31c82d0d0dde3f604a59eecfb3eed6c822579b68e07e6627a WatchSource:0}: Error finding container 66a3b53c0e30a7b31c82d0d0dde3f604a59eecfb3eed6c822579b68e07e6627a: Status 404 returned error can't find the container with id 66a3b53c0e30a7b31c82d0d0dde3f604a59eecfb3eed6c822579b68e07e6627a Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.651368 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.651558 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.151526904 +0000 UTC m=+163.324840266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.651646 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.652039 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.152023595 +0000 UTC m=+163.325336947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.655302 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8ae559a_06eb_4580_b4a8_c453e4e193a7.slice/crio-c313f0c2c7d636f8bdda087fd13268a177679a65195a226be63d863b30541fd2 WatchSource:0}: Error finding container c313f0c2c7d636f8bdda087fd13268a177679a65195a226be63d863b30541fd2: Status 404 returned error can't find the container with id c313f0c2c7d636f8bdda087fd13268a177679a65195a226be63d863b30541fd2 Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.664254 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cca1787_b4dd_4b49_a610_3effd6117c44.slice/crio-89cd283b33a8adec43efadd5c65ebb1e5a6e82aef6522155a9456b9ddde32f6f WatchSource:0}: Error finding container 89cd283b33a8adec43efadd5c65ebb1e5a6e82aef6522155a9456b9ddde32f6f: Status 404 returned error can't find the container with id 89cd283b33a8adec43efadd5c65ebb1e5a6e82aef6522155a9456b9ddde32f6f Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.674328 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f67f7b2_5a7d_421a_aa4a_333ef6d1e451.slice/crio-8dddd152dbf032b88182c5579c7a766e07ef460a00da1893dd42a3fb7bbd17f6 WatchSource:0}: Error finding container 8dddd152dbf032b88182c5579c7a766e07ef460a00da1893dd42a3fb7bbd17f6: Status 404 returned error can't find the container with id 8dddd152dbf032b88182c5579c7a766e07ef460a00da1893dd42a3fb7bbd17f6 Sep 29 10:24:52 crc kubenswrapper[4727]: W0929 10:24:52.687664 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod422a3020_82d4_48e2_ae59_9ab06061c5ee.slice/crio-8203e26cda945e6e27b34f5f50df43d94d49e9a597164b6ca0961f4772e09fd2 WatchSource:0}: Error finding container 8203e26cda945e6e27b34f5f50df43d94d49e9a597164b6ca0961f4772e09fd2: Status 404 returned error can't find the container with id 8203e26cda945e6e27b34f5f50df43d94d49e9a597164b6ca0961f4772e09fd2 Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.755511 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.755760 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.25571309 +0000 UTC m=+163.429026462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.755918 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.756433 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.256421546 +0000 UTC m=+163.429734918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.857059 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.857377 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.357319025 +0000 UTC m=+163.530632387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.857630 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.858162 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.358138935 +0000 UTC m=+163.531452387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.959216 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.959511 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.459461564 +0000 UTC m=+163.632774966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:52 crc kubenswrapper[4727]: I0929 10:24:52.959690 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:52 crc kubenswrapper[4727]: E0929 10:24:52.960010 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.459997116 +0000 UTC m=+163.633310478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.066137 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.076917 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.576763937 +0000 UTC m=+163.750077319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.169029 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.169647 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.669622448 +0000 UTC m=+163.842935850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.251830 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" event={"ID":"25b90489-d077-4da1-ae20-eb826bb3b189","Type":"ContainerStarted","Data":"7f4b3fd9e256c48e3e3d110dc0b76fcea4a2983d61479f2332be5c36d208b5f5"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.253746 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" event={"ID":"e00fd3e3-7070-497e-8370-ba2d427981dd","Type":"ContainerStarted","Data":"c1b8dd8eef0072bfcdebf69908cd6ce51f5feb89cb47ddf62f2066d50d01bb53"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.258733 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" event={"ID":"a8ae559a-06eb-4580-b4a8-c453e4e193a7","Type":"ContainerStarted","Data":"c313f0c2c7d636f8bdda087fd13268a177679a65195a226be63d863b30541fd2"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.263617 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" event={"ID":"39634023-d2e6-4b7e-a258-1d763c0bdaad","Type":"ContainerStarted","Data":"66a3b53c0e30a7b31c82d0d0dde3f604a59eecfb3eed6c822579b68e07e6627a"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.266518 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fkfsn" event={"ID":"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5","Type":"ContainerStarted","Data":"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.268174 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-r9ffv" event={"ID":"8cca1787-b4dd-4b49-a610-3effd6117c44","Type":"ContainerStarted","Data":"89cd283b33a8adec43efadd5c65ebb1e5a6e82aef6522155a9456b9ddde32f6f"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.269628 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" event={"ID":"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee","Type":"ContainerStarted","Data":"7dc648bf9383f2f76975724e29940eb345fd50e437a7c727f196c3919a49f6f2"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.271159 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.271468 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.771419198 +0000 UTC m=+163.944732600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.271756 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.272470 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.772449532 +0000 UTC m=+163.945763064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.273394 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2qq2r" event={"ID":"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451","Type":"ContainerStarted","Data":"8dddd152dbf032b88182c5579c7a766e07ef460a00da1893dd42a3fb7bbd17f6"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.275467 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" event={"ID":"9e431cd8-2196-4aed-a4f6-0505374c7561","Type":"ContainerStarted","Data":"65549995d1e8c0b2ac6a97837c6baa7640e49b3427ae3709c051ef6c2e8a95ec"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.277679 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" event={"ID":"422a3020-82d4-48e2-ae59-9ab06061c5ee","Type":"ContainerStarted","Data":"8203e26cda945e6e27b34f5f50df43d94d49e9a597164b6ca0961f4772e09fd2"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.281103 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" event={"ID":"a2ce6052-f1b1-4f55-98f7-7e509ae66846","Type":"ContainerStarted","Data":"9a4044e618f09a985b76fa0491d9c1af80fbf385a950267102c52265bc65021b"} Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.281216 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.287606 4727 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nxzg5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.287680 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.317411 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" podStartSLOduration=136.317382733 podStartE2EDuration="2m16.317382733s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:53.313373329 +0000 UTC m=+163.486686691" watchObservedRunningTime="2025-09-29 10:24:53.317382733 +0000 UTC m=+163.490696095" Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.373700 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.373907 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.873870804 +0000 UTC m=+164.047184176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.374217 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.375923 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.875895561 +0000 UTC m=+164.049209103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.475405 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.475593 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.975571102 +0000 UTC m=+164.148884464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.475660 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.476254 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:53.976209287 +0000 UTC m=+164.149522659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.576580 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.576765 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.076731427 +0000 UTC m=+164.250044819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.577070 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.577397 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.077384873 +0000 UTC m=+164.250698235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.678523 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.678658 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.17863757 +0000 UTC m=+164.351950932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.679017 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.679325 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.179316186 +0000 UTC m=+164.352629548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.781177 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.781382 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.28132681 +0000 UTC m=+164.454640172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.781660 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.782670 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.282488427 +0000 UTC m=+164.455801939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.882841 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.883041 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.383008758 +0000 UTC m=+164.556322130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.883767 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.884179 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.384160925 +0000 UTC m=+164.557474317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:53 crc kubenswrapper[4727]: I0929 10:24:53.984522 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:53 crc kubenswrapper[4727]: E0929 10:24:53.985134 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.485110475 +0000 UTC m=+164.658423857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.086625 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.087197 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.587171312 +0000 UTC m=+164.760484684 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.187192 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.187373 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.687321923 +0000 UTC m=+164.860635285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.187586 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.187902 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.687889127 +0000 UTC m=+164.861202489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.284708 4727 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nxzg5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.284766 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.287998 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.288489 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.788471179 +0000 UTC m=+164.961784541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.325273 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" podStartSLOduration=137.325253839 podStartE2EDuration="2m17.325253839s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:54.324175463 +0000 UTC m=+164.497488835" watchObservedRunningTime="2025-09-29 10:24:54.325253839 +0000 UTC m=+164.498567211" Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.389447 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.389830 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.889814918 +0000 UTC m=+165.063128280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.491952 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.492234 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.992195262 +0000 UTC m=+165.165508664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.492648 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.492955 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:54.99294337 +0000 UTC m=+165.166256732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.593558 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.593812 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.093773257 +0000 UTC m=+165.267086659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.594161 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.594718 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.094694489 +0000 UTC m=+165.268007901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.695245 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.695469 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.195439205 +0000 UTC m=+165.368752577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.696092 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.696558 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.19653802 +0000 UTC m=+165.369851412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.797491 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.797873 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.297826509 +0000 UTC m=+165.471139881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.797961 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.798381 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.29832217 +0000 UTC m=+165.471635742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.899102 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.899296 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.39926173 +0000 UTC m=+165.572575102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:54 crc kubenswrapper[4727]: I0929 10:24:54.899551 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:54 crc kubenswrapper[4727]: E0929 10:24:54.899861 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.399849104 +0000 UTC m=+165.573162466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.000906 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.001385 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.501294296 +0000 UTC m=+165.674607718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.001623 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.002219 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.502195597 +0000 UTC m=+165.675508989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.103547 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.104066 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.604026638 +0000 UTC m=+165.777340030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.104148 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.105068 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.605026172 +0000 UTC m=+165.778339584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.206096 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.206441 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.706383962 +0000 UTC m=+165.879697334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.206566 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.207219 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.707203561 +0000 UTC m=+165.880517123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.311194 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.311490 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.811454829 +0000 UTC m=+165.984768191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.311648 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.312175 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.812166375 +0000 UTC m=+165.985479737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.413372 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.413702 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.913653188 +0000 UTC m=+166.086966560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.413802 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.414835 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:55.914813735 +0000 UTC m=+166.088127107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.515802 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.516197 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.016140085 +0000 UTC m=+166.189453487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.516286 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.517440 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.017420335 +0000 UTC m=+166.190733727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.618647 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.619070 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.11900387 +0000 UTC m=+166.292317282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.619384 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.620015 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.119984633 +0000 UTC m=+166.293298035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.721525 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.721724 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.221699991 +0000 UTC m=+166.395013353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.721956 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.722268 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.222260434 +0000 UTC m=+166.395573796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.823800 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.823921 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.323898541 +0000 UTC m=+166.497211913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.824165 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.824479 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.324468344 +0000 UTC m=+166.497781706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.926254 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.926575 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.426537671 +0000 UTC m=+166.599851063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:55 crc kubenswrapper[4727]: I0929 10:24:55.927085 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:55 crc kubenswrapper[4727]: E0929 10:24:55.927556 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.427536914 +0000 UTC m=+166.600850316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.029139 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.029385 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.529319014 +0000 UTC m=+166.702632376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.029646 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.030127 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.530113203 +0000 UTC m=+166.703426755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.131113 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.131317 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.631287228 +0000 UTC m=+166.804600590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.131855 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.132231 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.63222289 +0000 UTC m=+166.805536252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.233075 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.233240 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.733219522 +0000 UTC m=+166.906532884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.233437 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.233929 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.733916938 +0000 UTC m=+166.907230300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.296093 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.299446 4727 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fcr2d container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.299566 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.324089 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" podStartSLOduration=138.324062976 podStartE2EDuration="2m18.324062976s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:56.322405477 +0000 UTC m=+166.495718839" watchObservedRunningTime="2025-09-29 10:24:56.324062976 +0000 UTC m=+166.497376338" Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.334430 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.334548 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.834527641 +0000 UTC m=+167.007841003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.334881 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.335217 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.835208347 +0000 UTC m=+167.008521709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.436958 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.437290 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.937229962 +0000 UTC m=+167.110543324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.437654 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.438214 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:56.938184884 +0000 UTC m=+167.111498286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.539648 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.539897 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.039864542 +0000 UTC m=+167.213177914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.540612 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.541296 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.041264315 +0000 UTC m=+167.214577717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.642378 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.642643 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.142601894 +0000 UTC m=+167.315915286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.642896 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.643279 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.14326539 +0000 UTC m=+167.316578762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.743762 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.743918 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.243898503 +0000 UTC m=+167.417211865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.744039 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.744331 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.244322643 +0000 UTC m=+167.417636005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.845469 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.845653 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.345619191 +0000 UTC m=+167.518932553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.846044 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.846693 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.346668566 +0000 UTC m=+167.519981968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.947555 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.947994 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.447968215 +0000 UTC m=+167.621281587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:56 crc kubenswrapper[4727]: I0929 10:24:56.948075 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:56 crc kubenswrapper[4727]: E0929 10:24:56.948463 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.448452626 +0000 UTC m=+167.621765998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.049401 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.049687 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.549642652 +0000 UTC m=+167.722956024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.050068 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.050687 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.550662726 +0000 UTC m=+167.723976088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.151567 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.151762 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.651733309 +0000 UTC m=+167.825046671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.152074 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.152558 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.652544108 +0000 UTC m=+167.825857480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.252951 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.253451 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.753394796 +0000 UTC m=+167.926708198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.302242 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" event={"ID":"94497a6a-de54-46ae-a609-12872ceb5b24","Type":"ContainerStarted","Data":"b447791c35187305c9c2201a4b4c660c1825b6da9a2bd3f0dd26016b707e3f8b"} Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.304114 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" event={"ID":"61d8d661-e0b8-47ff-8733-08c97804c66f","Type":"ContainerStarted","Data":"c4a76b594918510389c2710f9c624d32ca068ba88b244c10c6197027ee6f83e4"} Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.306632 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" event={"ID":"fab27910-938e-4922-b9b0-b2cf44306cd6","Type":"ContainerStarted","Data":"7c12e9ec0624afe61d90a27edc8059764ed591d352f6071487d02c7508b485ed"} Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.308205 4727 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fcr2d container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.308294 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.354691 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.355224 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.855200897 +0000 UTC m=+168.028514359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.455929 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.456174 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.956141196 +0000 UTC m=+168.129454598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.456444 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.457496 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:57.957474707 +0000 UTC m=+168.130788109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.557728 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.558429 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.058401467 +0000 UTC m=+168.231714829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.659722 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.659999 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.159988163 +0000 UTC m=+168.333301525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.760638 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.761197 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.261167959 +0000 UTC m=+168.434481331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.762287 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.762948 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.26293437 +0000 UTC m=+168.436247732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.862917 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.863431 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.363412249 +0000 UTC m=+168.536725621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:57 crc kubenswrapper[4727]: I0929 10:24:57.964564 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:57 crc kubenswrapper[4727]: E0929 10:24:57.965229 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.465196399 +0000 UTC m=+168.638509821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.065517 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.066097 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.566081638 +0000 UTC m=+168.739395000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.167740 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.168028 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.668016192 +0000 UTC m=+168.841329554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.269260 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.269521 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.769474974 +0000 UTC m=+168.942788376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.269636 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.270195 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.77017345 +0000 UTC m=+168.943486842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.314569 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" event={"ID":"64c27281-4623-451c-84b9-48b20e3a3a7d","Type":"ContainerStarted","Data":"0365a72e358ffdfd0cdcc2998279cccaf3ee29d7b47a7e56cb57436bfa9f9332"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.315967 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" event={"ID":"73a9e11c-a13e-4167-9744-d7427e583d89","Type":"ContainerStarted","Data":"2fc04b4f148353e3533d18e40f8e74abef9fd72a0a434d66b1f61813c57a45d1"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.317454 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" event={"ID":"79209de4-57ff-43d1-9f46-aae7501d21c6","Type":"ContainerStarted","Data":"b2b620989db940e9ebe37fd1470e042f795b76e9362d7d8ebe8e52178261f057"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.319857 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" event={"ID":"6404327b-095b-435a-b953-b1c3b7a6f332","Type":"ContainerStarted","Data":"276c3a00185cbe28110b11b1f8922ba9c2107476ef9036db2eecd77117243780"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.321238 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" event={"ID":"305b863d-a41d-437c-986b-04a2bb52fde9","Type":"ContainerStarted","Data":"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.322698 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" event={"ID":"88be3001-6d99-49d8-b994-90178f3e64e0","Type":"ContainerStarted","Data":"1c0f371c67a0759d2cce9b20b7fa24b597d55d7c05b183d3906e08c357beedd1"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.323972 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t4lhc" event={"ID":"274d4012-934c-426c-8902-20d19732cfea","Type":"ContainerStarted","Data":"516fad543c3fe06d69b27ecca67b650dbcb5f5008a908095f2e4a91001cf2d65"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.325073 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" event={"ID":"01e7cad1-1c3e-452e-979c-3ff3787050d1","Type":"ContainerStarted","Data":"7c4e733c3ea4e267031a7523229add9a8eb15fc2140c205fb36a22e2b37eb46a"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.326193 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" event={"ID":"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44","Type":"ContainerStarted","Data":"84d4b21d4c26906013546a6db1a2fee19b5de2961597bdde7a98ae2642f084bf"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.327603 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" event={"ID":"ebca6c99-76f7-4e93-9d6c-a91566decffc","Type":"ContainerStarted","Data":"fc4421b26c1f494864f74c158c6a364b992125dc209e1cc8ca4f7b7db5ed4ca5"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.329190 4727 generic.go:334] "Generic (PLEG): container finished" podID="04eb68eb-eab0-4868-96e6-38781e313d3a" containerID="71446c25c015514b9ffcf276affd5a5e6263969780b5ed1616366e07b378a82d" exitCode=0 Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.329233 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" event={"ID":"04eb68eb-eab0-4868-96e6-38781e313d3a","Type":"ContainerDied","Data":"71446c25c015514b9ffcf276affd5a5e6263969780b5ed1616366e07b378a82d"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.330949 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" event={"ID":"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c","Type":"ContainerStarted","Data":"163fc16dc56b4e35c99521b3ba83c0fdeba68c7378ddd93ee714819ca5fba90b"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.333053 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" event={"ID":"46124f8a-3360-4dfe-b468-5addce47a614","Type":"ContainerStarted","Data":"b51c56608b1342076b4b02729b489f4ddf9f253b1cf23d10121cb487cfdb58ca"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.335032 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" event={"ID":"23cd88cb-71a1-4e96-9091-152472b884a1","Type":"ContainerStarted","Data":"db637e61a0c4857456456eb32546bda0aa4d0e9bb081294c09f6a70d0e23ac74"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.337538 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" event={"ID":"f65657cc-1eae-43a4-82e2-65379bbed510","Type":"ContainerStarted","Data":"e0d1bef37cff55648762b9f4b1e2eaa3c9cb4aec72ef97efc900690572e45765"} Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.359967 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-fkfsn" podStartSLOduration=141.35994991 podStartE2EDuration="2m21.35994991s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:58.35741561 +0000 UTC m=+168.530729012" watchObservedRunningTime="2025-09-29 10:24:58.35994991 +0000 UTC m=+168.533263272" Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.371064 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.371245 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.871222433 +0000 UTC m=+169.044535795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.371696 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.372171 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.872146125 +0000 UTC m=+169.045459517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.472798 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.472976 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.972952512 +0000 UTC m=+169.146265874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.473086 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.473535 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:58.973520045 +0000 UTC m=+169.146833407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.573953 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.574164 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.074134768 +0000 UTC m=+169.247448130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.574968 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.575325 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.075316475 +0000 UTC m=+169.248629837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.675948 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.676093 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.176072571 +0000 UTC m=+169.349385943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.676377 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.676774 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.176763948 +0000 UTC m=+169.350077320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.777780 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.777981 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.277951063 +0000 UTC m=+169.451264425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.778430 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.778739 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.278732052 +0000 UTC m=+169.452045414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.879493 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.879651 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.379622921 +0000 UTC m=+169.552936283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.879854 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.880349 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.380322637 +0000 UTC m=+169.553635999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:58 crc kubenswrapper[4727]: I0929 10:24:58.981317 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:58 crc kubenswrapper[4727]: E0929 10:24:58.981746 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.481731528 +0000 UTC m=+169.655044890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.082967 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.083250 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.583239622 +0000 UTC m=+169.756552984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.184177 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.184414 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.684392957 +0000 UTC m=+169.857706319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.184864 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.185229 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.685221037 +0000 UTC m=+169.858534389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.286040 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.286184 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.786167457 +0000 UTC m=+169.959480819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.286363 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.286762 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.786746131 +0000 UTC m=+169.960059493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.347894 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" event={"ID":"25b90489-d077-4da1-ae20-eb826bb3b189","Type":"ContainerStarted","Data":"1f736aab8d5c9358f4d8aac440c45351976c4425d25aedf8efafa0282d3544a1"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.351769 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" event={"ID":"d6c27c2f-cae2-40bc-8889-9a14daedc8ed","Type":"ContainerStarted","Data":"6abb85a4905fad7164f4ad0e1f4d8bf70e3b3a4cdd478c98914bd409636d76a5"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.353509 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zg47h" event={"ID":"65081bdc-ce5d-4a32-8c55-91316d66a83a","Type":"ContainerStarted","Data":"71de028b27a6c80b084402259659148c4a6d6f6e486d9b656c4f490b7b781c78"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.354918 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" event={"ID":"e00fd3e3-7070-497e-8370-ba2d427981dd","Type":"ContainerStarted","Data":"6467bdd54391a52fad476702356c7fcffa7ece502b5b99baadcf644e8efedc64"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.356410 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2qq2r" event={"ID":"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451","Type":"ContainerStarted","Data":"7f2150e57aec7ee75945821e9672c7989096aa20803a8ed93edf792ad10cc252"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.357898 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-r9ffv" event={"ID":"8cca1787-b4dd-4b49-a610-3effd6117c44","Type":"ContainerStarted","Data":"391d7a1fee43505d569f2b4b3f4abecf803eff08ec7ae0a84e8d3fa92409b842"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.359802 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lljzg" event={"ID":"dc06dadc-ca42-476d-ac08-656b17e587e8","Type":"ContainerStarted","Data":"27b4395b2bac428af92884a9fadf1828ea015c569334467fca534615f6c7a2d3"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.362163 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" event={"ID":"d6a94e2a-c1ed-43c6-91d9-495a0bb673ce","Type":"ContainerStarted","Data":"765ccb5fa689270bb6df99020de2a1cabba074d4f25378f4181a5e3f8339da53"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.364274 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" event={"ID":"a8ae559a-06eb-4580-b4a8-c453e4e193a7","Type":"ContainerStarted","Data":"efd2d28c32c67eb89ee7caca532490d023e95d6db02572aa4b314e13c93b2fdf"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.367909 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" event={"ID":"9e431cd8-2196-4aed-a4f6-0505374c7561","Type":"ContainerStarted","Data":"445722b4898e25960991e170332dcd5e220625c2ca04e262d77a4ceedf2d089c"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.369630 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" event={"ID":"53f5f641-e101-4fa8-941e-540282897459","Type":"ContainerStarted","Data":"c1464cf53b46809300aedc5c51689f9b8518fbd40ab5722664f566d671ed7051"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.372974 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" event={"ID":"c535a2d5-d067-4fd7-84d7-419e456bebfa","Type":"ContainerStarted","Data":"d4ce525aa880ade02a94fd1a31c18e7ef349378b243e920c7f0d1aaeac80eecf"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.374305 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" event={"ID":"1030d2bc-8c0b-42b6-88da-99ca8f63a1ff","Type":"ContainerStarted","Data":"817b54e039db3c1d4a504f72bf0ea7d1857b63654cc59c27d56d8d16d443a68a"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.375405 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" event={"ID":"a2ce6052-f1b1-4f55-98f7-7e509ae66846","Type":"ContainerStarted","Data":"2dec00dd3219ca8cfda1430bd21f4ebfb9589febeec76dcaaf80193d85c459ca"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.376215 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" event={"ID":"422a3020-82d4-48e2-ae59-9ab06061c5ee","Type":"ContainerStarted","Data":"72a011bd0f82d754e218170c27f3134ec3c3de2a539f5733427bbfed59631008"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.377145 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" event={"ID":"39634023-d2e6-4b7e-a258-1d763c0bdaad","Type":"ContainerStarted","Data":"8d734023231610ec79f3c431cc2ab35340e028240f29f57545c9a9d52d6af507"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.378205 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" event={"ID":"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2","Type":"ContainerStarted","Data":"ec3c7e62a378703afcf313eb626eaecf2011a5dcc0c5e2cbe9fc52b0bd80ecdc"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.380106 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" event={"ID":"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a","Type":"ContainerStarted","Data":"5b6a212b2bfac04c8c6334fa9ed10b6b71372c08074f04155c5f0310b1bbc0ac"} Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.380154 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382726 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382757 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382916 4727 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-65ccr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382944 4727 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vvk6t container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382960 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.382990 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" podUID="f65657cc-1eae-43a4-82e2-65379bbed510" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.385233 4727 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r97m9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.385279 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.388869 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.389417 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:24:59.889400941 +0000 UTC m=+170.062714303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.391565 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rz2nt" podStartSLOduration=141.391554231 podStartE2EDuration="2m21.391554231s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.389659707 +0000 UTC m=+169.562973069" watchObservedRunningTime="2025-09-29 10:24:59.391554231 +0000 UTC m=+169.564867593" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.403839 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" podStartSLOduration=142.403822018 podStartE2EDuration="2m22.403822018s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.403443869 +0000 UTC m=+169.576757251" watchObservedRunningTime="2025-09-29 10:24:59.403822018 +0000 UTC m=+169.577135380" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.417372 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nw5vx" podStartSLOduration=142.417354274 podStartE2EDuration="2m22.417354274s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.41674378 +0000 UTC m=+169.590057142" watchObservedRunningTime="2025-09-29 10:24:59.417354274 +0000 UTC m=+169.590667636" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.439505 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f7hn5" podStartSLOduration=141.439484842 podStartE2EDuration="2m21.439484842s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.431847863 +0000 UTC m=+169.605161245" watchObservedRunningTime="2025-09-29 10:24:59.439484842 +0000 UTC m=+169.612798204" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.464521 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vgphv" podStartSLOduration=142.464505717 podStartE2EDuration="2m22.464505717s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.462524701 +0000 UTC m=+169.635838063" watchObservedRunningTime="2025-09-29 10:24:59.464505717 +0000 UTC m=+169.637819079" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.479949 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.486844 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.495064 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.495799 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-t4lhc" podStartSLOduration=142.495789018 podStartE2EDuration="2m22.495789018s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.494809886 +0000 UTC m=+169.668123248" watchObservedRunningTime="2025-09-29 10:24:59.495789018 +0000 UTC m=+169.669102380" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.504061 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.004046362 +0000 UTC m=+170.177359724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.521515 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.521590 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.530110 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-q6kbd" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.531217 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7ws9x" podStartSLOduration=141.531198616 podStartE2EDuration="2m21.531198616s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.525230967 +0000 UTC m=+169.698544329" watchObservedRunningTime="2025-09-29 10:24:59.531198616 +0000 UTC m=+169.704511978" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.547609 4727 patch_prober.go:28] interesting pod/console-f9d7485db-fkfsn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.547678 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fkfsn" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.554048 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" podStartSLOduration=141.55402033 podStartE2EDuration="2m21.55402033s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.548097572 +0000 UTC m=+169.721410934" watchObservedRunningTime="2025-09-29 10:24:59.55402033 +0000 UTC m=+169.727333692" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.598043 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.601471 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.101450699 +0000 UTC m=+170.274764061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.612358 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-h9nz2" podStartSLOduration=142.612305613 podStartE2EDuration="2m22.612305613s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.606943567 +0000 UTC m=+169.780256929" watchObservedRunningTime="2025-09-29 10:24:59.612305613 +0000 UTC m=+169.785618975" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.636122 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thw88" podStartSLOduration=142.636107539 podStartE2EDuration="2m22.636107539s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.633224342 +0000 UTC m=+169.806537704" watchObservedRunningTime="2025-09-29 10:24:59.636107539 +0000 UTC m=+169.809420901" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.660431 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.674896 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m4jkd" podStartSLOduration=142.674876526 podStartE2EDuration="2m22.674876526s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.672670124 +0000 UTC m=+169.845983486" watchObservedRunningTime="2025-09-29 10:24:59.674876526 +0000 UTC m=+169.848189898" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.685876 4727 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r97m9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.685928 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.711871 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.713949 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.213928379 +0000 UTC m=+170.387241911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.720921 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jc8lh" podStartSLOduration=142.720898082 podStartE2EDuration="2m22.720898082s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.720842791 +0000 UTC m=+169.894156153" watchObservedRunningTime="2025-09-29 10:24:59.720898082 +0000 UTC m=+169.894211454" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.759920 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vm4v4" podStartSLOduration=142.759892524 podStartE2EDuration="2m22.759892524s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.75888639 +0000 UTC m=+169.932199752" watchObservedRunningTime="2025-09-29 10:24:59.759892524 +0000 UTC m=+169.933205886" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.813982 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.815410 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.315386741 +0000 UTC m=+170.488700103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.827799 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podStartSLOduration=141.827774741 podStartE2EDuration="2m21.827774741s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:24:59.804463716 +0000 UTC m=+169.977777078" watchObservedRunningTime="2025-09-29 10:24:59.827774741 +0000 UTC m=+170.001088103" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.916202 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:24:59 crc kubenswrapper[4727]: E0929 10:24:59.916632 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.416619279 +0000 UTC m=+170.589932641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.981973 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.982094 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.983752 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.983819 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.994637 4727 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-65ccr container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.994721 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.994876 4727 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-65ccr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Sep 29 10:24:59 crc kubenswrapper[4727]: I0929 10:24:59.994935 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.012783 4727 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vvk6t container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.012839 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" podUID="f65657cc-1eae-43a4-82e2-65379bbed510" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.012938 4727 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vvk6t container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.013042 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" podUID="f65657cc-1eae-43a4-82e2-65379bbed510" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.017522 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.017686 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.517658601 +0000 UTC m=+170.690971963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.018057 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.018564 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.518549012 +0000 UTC m=+170.691862384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.118899 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.119067 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.619035592 +0000 UTC m=+170.792348974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.119106 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.119486 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.619474982 +0000 UTC m=+170.792788364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.220376 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.220579 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.720559286 +0000 UTC m=+170.893872648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.220622 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.220963 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.720947765 +0000 UTC m=+170.894261127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.321642 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.321899 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.821851644 +0000 UTC m=+170.995165006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.322201 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.322740 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.822718904 +0000 UTC m=+170.996032276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.384389 4727 generic.go:334] "Generic (PLEG): container finished" podID="422a3020-82d4-48e2-ae59-9ab06061c5ee" containerID="72a011bd0f82d754e218170c27f3134ec3c3de2a539f5733427bbfed59631008" exitCode=0 Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.384525 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" event={"ID":"422a3020-82d4-48e2-ae59-9ab06061c5ee","Type":"ContainerDied","Data":"72a011bd0f82d754e218170c27f3134ec3c3de2a539f5733427bbfed59631008"} Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.385058 4727 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r97m9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.385063 4727 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-65ccr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.385099 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.385131 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.385424 4727 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vvk6t container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.386030 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" podUID="f65657cc-1eae-43a4-82e2-65379bbed510" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.424010 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.424209 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.924181167 +0000 UTC m=+171.097494529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.424412 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.424801 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:00.924786081 +0000 UTC m=+171.098099443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.433304 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l7ggf" podStartSLOduration=143.433261779 podStartE2EDuration="2m23.433261779s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:00.430252139 +0000 UTC m=+170.603565511" watchObservedRunningTime="2025-09-29 10:25:00.433261779 +0000 UTC m=+170.606575141" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.460845 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" podStartSLOduration=142.460822213 podStartE2EDuration="2m22.460822213s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:00.458177432 +0000 UTC m=+170.631490794" watchObservedRunningTime="2025-09-29 10:25:00.460822213 +0000 UTC m=+170.634135575" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.477026 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zg47h" podStartSLOduration=13.477010142 podStartE2EDuration="13.477010142s" podCreationTimestamp="2025-09-29 10:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:00.476305495 +0000 UTC m=+170.649618847" watchObservedRunningTime="2025-09-29 10:25:00.477010142 +0000 UTC m=+170.650323504" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.489274 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pp54r" podStartSLOduration=143.489259088 podStartE2EDuration="2m23.489259088s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:00.486694268 +0000 UTC m=+170.660007630" watchObservedRunningTime="2025-09-29 10:25:00.489259088 +0000 UTC m=+170.662572450" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.525722 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.525933 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.025903315 +0000 UTC m=+171.199216677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.526823 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.529793 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.029780276 +0000 UTC m=+171.203093638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.627921 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.628203 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.128177897 +0000 UTC m=+171.301491279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.628303 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.628384 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.628912 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.128899024 +0000 UTC m=+171.302212386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.645419 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59055eb3-6cad-4335-b100-39955f1c0500-metrics-certs\") pod \"network-metrics-daemon-gttbj\" (UID: \"59055eb3-6cad-4335-b100-39955f1c0500\") " pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.651997 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gttbj" Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.729952 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.730761 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.230746215 +0000 UTC m=+171.404059577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.834761 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.835332 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.33531164 +0000 UTC m=+171.508625002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.881146 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gttbj"] Sep 29 10:25:00 crc kubenswrapper[4727]: W0929 10:25:00.890888 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59055eb3_6cad_4335_b100_39955f1c0500.slice/crio-f65db28a7ecf973854541531d3f534910092c7c583625bfc7030d64a049592b3 WatchSource:0}: Error finding container f65db28a7ecf973854541531d3f534910092c7c583625bfc7030d64a049592b3: Status 404 returned error can't find the container with id f65db28a7ecf973854541531d3f534910092c7c583625bfc7030d64a049592b3 Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.935984 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.936389 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.436359793 +0000 UTC m=+171.609673155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.936452 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:00 crc kubenswrapper[4727]: E0929 10:25:00.936913 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.436894824 +0000 UTC m=+171.610208346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.984983 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Sep 29 10:25:00 crc kubenswrapper[4727]: I0929 10:25:00.985594 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.039154 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.039681 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.539655107 +0000 UTC m=+171.712968469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.140720 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.141006 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.640994617 +0000 UTC m=+171.814307979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.242220 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.242433 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.742415458 +0000 UTC m=+171.915728820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.242648 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.242942 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.74293293 +0000 UTC m=+171.916246292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.371394 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.371814 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.871762173 +0000 UTC m=+172.045075565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.372053 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.372581 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.872565471 +0000 UTC m=+172.045878853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.392574 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" event={"ID":"a49f0f4d-7c88-4f1a-b2d3-f274f4d0fd44","Type":"ContainerStarted","Data":"62d1ae0da3b5fcc3064343cf84caf5ca8e398ab2ec3552a4228b7d827925c9f6"} Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.395729 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" event={"ID":"ebca6c99-76f7-4e93-9d6c-a91566decffc","Type":"ContainerStarted","Data":"7a5e88e916192cf63fb9c4cb1da5270342fe6610741eec6b83afa45b84a16734"} Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.397638 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" event={"ID":"194f9707-e39e-4a4a-bda7-c6c69b0b2e7c","Type":"ContainerStarted","Data":"ca358ff3fe76e69e838afe6c821d435470aab74d9ffa6ff18c0ca25f67845613"} Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.400202 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" event={"ID":"6404327b-095b-435a-b953-b1c3b7a6f332","Type":"ContainerStarted","Data":"1e69001276c4be9e34acaba1b1644cd358ac7ea5dc556c03ec7cc96023e9e4fd"} Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.403500 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gttbj" event={"ID":"59055eb3-6cad-4335-b100-39955f1c0500","Type":"ContainerStarted","Data":"f65db28a7ecf973854541531d3f534910092c7c583625bfc7030d64a049592b3"} Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.403548 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.404879 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.410802 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.410851 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.410857 4727 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-x7jzw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.410913 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" podUID="9e431cd8-2196-4aed-a4f6-0505374c7561" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.427309 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-r9ffv" podStartSLOduration=14.427287721 podStartE2EDuration="14.427287721s" podCreationTimestamp="2025-09-29 10:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.42597079 +0000 UTC m=+171.599284152" watchObservedRunningTime="2025-09-29 10:25:01.427287721 +0000 UTC m=+171.600601093" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.469567 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" podStartSLOduration=143.469536169 podStartE2EDuration="2m23.469536169s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.465409822 +0000 UTC m=+171.638723194" watchObservedRunningTime="2025-09-29 10:25:01.469536169 +0000 UTC m=+171.642849551" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.473082 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.473304 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.973267966 +0000 UTC m=+172.146581348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.474419 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.475163 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:01.97514871 +0000 UTC m=+172.148462072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.529322 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" podStartSLOduration=144.529298336 podStartE2EDuration="2m24.529298336s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.515636067 +0000 UTC m=+171.688949429" watchObservedRunningTime="2025-09-29 10:25:01.529298336 +0000 UTC m=+171.702611698" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.530679 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2qq2r" podStartSLOduration=144.530671968 podStartE2EDuration="2m24.530671968s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.49309455 +0000 UTC m=+171.666407932" watchObservedRunningTime="2025-09-29 10:25:01.530671968 +0000 UTC m=+171.703985340" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.537513 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vl9k7" podStartSLOduration=143.537494838 podStartE2EDuration="2m23.537494838s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.536797622 +0000 UTC m=+171.710110984" watchObservedRunningTime="2025-09-29 10:25:01.537494838 +0000 UTC m=+171.710808200" Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.576136 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.576317 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.076250894 +0000 UTC m=+172.249564256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.576502 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.577027 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.077018302 +0000 UTC m=+172.250331664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.678045 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.678466 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.178434173 +0000 UTC m=+172.351747535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.678774 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.679308 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.179282533 +0000 UTC m=+172.352596075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.779602 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.779769 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.279738252 +0000 UTC m=+172.453051614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.779814 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.780368 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.280323086 +0000 UTC m=+172.453636448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.880960 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.881253 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.381228425 +0000 UTC m=+172.554541787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.881813 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.882375 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.382354592 +0000 UTC m=+172.555667954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.983119 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.983412 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.483363123 +0000 UTC m=+172.656676495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.984015 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:01 crc kubenswrapper[4727]: E0929 10:25:01.984648 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.484636613 +0000 UTC m=+172.657950115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.987671 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:01 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:01 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:01 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:01 crc kubenswrapper[4727]: I0929 10:25:01.987761 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.084681 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.084885 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.584853067 +0000 UTC m=+172.758166429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.085321 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.085729 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.585712807 +0000 UTC m=+172.759026159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.186930 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.187279 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.687256671 +0000 UTC m=+172.860570033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.289036 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.289698 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.789671226 +0000 UTC m=+172.962984588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.397001 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.397308 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.897257401 +0000 UTC m=+173.070570773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.397644 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.398216 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.898195753 +0000 UTC m=+173.071509285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.412139 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lljzg" event={"ID":"dc06dadc-ca42-476d-ac08-656b17e587e8","Type":"ContainerStarted","Data":"e9ee49eba1ec6ed0da70a0ba01255a3817379944ffd27b32a5915ce727f94b6b"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.412371 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-lljzg" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.414067 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" event={"ID":"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee","Type":"ContainerStarted","Data":"35e0cfdf315ceb7834ca0b97894c081e1dd15b1f8c31ebd444872dc2a4106b4e"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.416308 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" event={"ID":"25b90489-d077-4da1-ae20-eb826bb3b189","Type":"ContainerStarted","Data":"e956fcb515dbaa298aedf47ba92333699ce92318d38beee371d26a91724d8810"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.420964 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" event={"ID":"e00fd3e3-7070-497e-8370-ba2d427981dd","Type":"ContainerStarted","Data":"9f2421403dfeb095ce63ee00e54e94da21ac31c6c98ec014f1058261bddb78ea"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.423782 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" event={"ID":"04eb68eb-eab0-4868-96e6-38781e313d3a","Type":"ContainerStarted","Data":"a72b45b73bfd50354af9f51f7efc660ed1a7547ee9ff9742b9331a43cca870ed"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.423833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" event={"ID":"04eb68eb-eab0-4868-96e6-38781e313d3a","Type":"ContainerStarted","Data":"3a4bb62229e87503b44ec1905990d8f67324afe5e9790fe137ed1ef5d79faf16"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.427788 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" event={"ID":"39634023-d2e6-4b7e-a258-1d763c0bdaad","Type":"ContainerStarted","Data":"f73037093d41b144668ebf17a4a36a71a38cfb7605f7e409714102f139b2b3ac"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.428631 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.433317 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" event={"ID":"9fe1e500-bf28-4968-bca6-4e5e53ee2a1a","Type":"ContainerStarted","Data":"773167b62e0efbf373d5add5ebed3522dd2e76da91fde5b3dad2cd1c008495c0"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.437981 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gttbj" event={"ID":"59055eb3-6cad-4335-b100-39955f1c0500","Type":"ContainerStarted","Data":"b5db6b23a0fab0ab3b7ab7e12bfdeff303b1b9423ad0073037e30c7e7a6a84fd"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.442076 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-lljzg" podStartSLOduration=15.442054399 podStartE2EDuration="15.442054399s" podCreationTimestamp="2025-09-29 10:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.440620415 +0000 UTC m=+172.613933787" watchObservedRunningTime="2025-09-29 10:25:02.442054399 +0000 UTC m=+172.615367761" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.443606 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" podStartSLOduration=144.443598535 podStartE2EDuration="2m24.443598535s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:01.55855414 +0000 UTC m=+171.731867522" watchObservedRunningTime="2025-09-29 10:25:02.443598535 +0000 UTC m=+172.616911897" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.446109 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" event={"ID":"53f5f641-e101-4fa8-941e-540282897459","Type":"ContainerStarted","Data":"b2ef55e39d137fc9db1be4aba1156326059d61410a4633e1b2fe7b29fa9ff58f"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.450621 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" event={"ID":"422a3020-82d4-48e2-ae59-9ab06061c5ee","Type":"ContainerStarted","Data":"f58de77b16d214ae50cb554342804397cad7b3baa753f0cf4f172be29aec8d47"} Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.451183 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.452789 4727 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-x7jzw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.452825 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" podUID="9e431cd8-2196-4aed-a4f6-0505374c7561" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.452914 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.452998 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.462261 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.463360 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.468383 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.468537 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.479535 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpkh2" podStartSLOduration=144.479503844 podStartE2EDuration="2m24.479503844s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.473506714 +0000 UTC m=+172.646820096" watchObservedRunningTime="2025-09-29 10:25:02.479503844 +0000 UTC m=+172.652817206" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.484017 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.498870 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.499030 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:02.99900264 +0000 UTC m=+173.172316002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.499377 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.499417 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.500038 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.500191 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2j69r" podStartSLOduration=144.500143147 podStartE2EDuration="2m24.500143147s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.497167758 +0000 UTC m=+172.670481120" watchObservedRunningTime="2025-09-29 10:25:02.500143147 +0000 UTC m=+172.673456509" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.514812 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.01479344 +0000 UTC m=+173.188106802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.520969 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" podStartSLOduration=144.520947793 podStartE2EDuration="2m24.520947793s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.520068293 +0000 UTC m=+172.693381655" watchObservedRunningTime="2025-09-29 10:25:02.520947793 +0000 UTC m=+172.694261155" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.557914 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" podStartSLOduration=145.557892427 podStartE2EDuration="2m25.557892427s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.555831859 +0000 UTC m=+172.729145221" watchObservedRunningTime="2025-09-29 10:25:02.557892427 +0000 UTC m=+172.731205789" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.606494 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-krfsh" podStartSLOduration=145.606452073 podStartE2EDuration="2m25.606452073s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.604029516 +0000 UTC m=+172.777342878" watchObservedRunningTime="2025-09-29 10:25:02.606452073 +0000 UTC m=+172.779765435" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.608597 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.608967 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.609040 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.609699 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.109678888 +0000 UTC m=+173.282992250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.609738 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.644656 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.668072 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" podStartSLOduration=145.668021173 podStartE2EDuration="2m25.668021173s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.656453112 +0000 UTC m=+172.829766474" watchObservedRunningTime="2025-09-29 10:25:02.668021173 +0000 UTC m=+172.841334535" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.669138 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mwh6n" podStartSLOduration=145.669130008 podStartE2EDuration="2m25.669130008s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.634793096 +0000 UTC m=+172.808106458" watchObservedRunningTime="2025-09-29 10:25:02.669130008 +0000 UTC m=+172.842443370" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.681693 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zn6h" podStartSLOduration=145.681670782 podStartE2EDuration="2m25.681670782s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.67990881 +0000 UTC m=+172.853222172" watchObservedRunningTime="2025-09-29 10:25:02.681670782 +0000 UTC m=+172.854984144" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.700230 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2jgn" podStartSLOduration=144.70020396500001 podStartE2EDuration="2m24.700203965s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.697811959 +0000 UTC m=+172.871125351" watchObservedRunningTime="2025-09-29 10:25:02.700203965 +0000 UTC m=+172.873517327" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.711359 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.712172 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.212151324 +0000 UTC m=+173.385464686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.731376 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-lw586" podStartSLOduration=144.731354533 podStartE2EDuration="2m24.731354533s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.728546068 +0000 UTC m=+172.901859430" watchObservedRunningTime="2025-09-29 10:25:02.731354533 +0000 UTC m=+172.904667895" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.764526 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-q4qs6" podStartSLOduration=144.764507219 podStartE2EDuration="2m24.764507219s" podCreationTimestamp="2025-09-29 10:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:02.764253723 +0000 UTC m=+172.937567085" watchObservedRunningTime="2025-09-29 10:25:02.764507219 +0000 UTC m=+172.937820581" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.785739 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.816851 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.817240 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.317225421 +0000 UTC m=+173.490538783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.920174 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:02 crc kubenswrapper[4727]: E0929 10:25:02.920758 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.420733642 +0000 UTC m=+173.594046994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.988130 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:02 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:02 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:02 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:02 crc kubenswrapper[4727]: I0929 10:25:02.988217 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.021849 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.022178 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.522133173 +0000 UTC m=+173.695446535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.022270 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.023101 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.523090735 +0000 UTC m=+173.696404097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.123509 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.124167 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.624139838 +0000 UTC m=+173.797453200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.225512 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.225818 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.725803595 +0000 UTC m=+173.899116957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.329758 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.330129 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.830114244 +0000 UTC m=+174.003427606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.431078 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.431865 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:03.931853053 +0000 UTC m=+174.105166415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.472913 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gttbj" event={"ID":"59055eb3-6cad-4335-b100-39955f1c0500","Type":"ContainerStarted","Data":"37d2ab2132ba7028de87a70ecf62f4d31c4a35f65ed9facbbad901de71811c49"} Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.533456 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.535202 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.035168519 +0000 UTC m=+174.208481911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.569224 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.636001 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.636480 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.136466197 +0000 UTC m=+174.309779559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.736882 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.737147 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.237117631 +0000 UTC m=+174.410430993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.737288 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.737959 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.23794417 +0000 UTC m=+174.411257532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.839374 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.839477 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.339457574 +0000 UTC m=+174.512770936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.839574 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.839918 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.339908804 +0000 UTC m=+174.513222166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.941456 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.941650 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.441615702 +0000 UTC m=+174.614929064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.941866 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:03 crc kubenswrapper[4727]: E0929 10:25:03.942330 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.442310829 +0000 UTC m=+174.615624191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.990986 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:03 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:03 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:03 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:03 crc kubenswrapper[4727]: I0929 10:25:03.991112 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.043364 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.043603 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.543563596 +0000 UTC m=+174.716876958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.043704 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.044100 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.544083688 +0000 UTC m=+174.717397050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.145148 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.145371 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.645325996 +0000 UTC m=+174.818639358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.145505 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.145867 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.645851158 +0000 UTC m=+174.819164520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.248185 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.248409 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.748376085 +0000 UTC m=+174.921689447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.248624 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.249032 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.74902192 +0000 UTC m=+174.922335282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.350492 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.351041 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.851011165 +0000 UTC m=+175.024324527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.452736 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.453113 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:04.953099032 +0000 UTC m=+175.126412394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.475374 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.475753 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.479111 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f5c419a2-0bf0-4646-9a7d-df769977d3ad","Type":"ContainerStarted","Data":"d800a459b0190ee67c5ebef557f6de972db383acf1ff7a1e5365ac22aecf2ffa"} Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.480316 4727 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-k7cws container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.480374 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" podUID="422a3020-82d4-48e2-ae59-9ab06061c5ee" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.554632 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.555962 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.055903415 +0000 UTC m=+175.229216777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.573785 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.597160 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gttbj" podStartSLOduration=147.597137289 podStartE2EDuration="2m27.597137289s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:03.570155637 +0000 UTC m=+173.743469009" watchObservedRunningTime="2025-09-29 10:25:04.597137289 +0000 UTC m=+174.770450651" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.636485 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.636752 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.639312 4727 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hr9hg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.639396 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" podUID="04eb68eb-eab0-4868-96e6-38781e313d3a" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.656613 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.657189 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.157157913 +0000 UTC m=+175.330471505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.758110 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.758433 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.25839204 +0000 UTC m=+175.431705412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.758842 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.760266 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.260250833 +0000 UTC m=+175.433564195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.861859 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.862115 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.362078404 +0000 UTC m=+175.535391776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.862379 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.862891 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.362872783 +0000 UTC m=+175.536186155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.963183 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.963392 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.463352632 +0000 UTC m=+175.636666004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.963732 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:04 crc kubenswrapper[4727]: E0929 10:25:04.964106 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.46409558 +0000 UTC m=+175.637408952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.986097 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:04 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:04 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:04 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:04 crc kubenswrapper[4727]: I0929 10:25:04.986192 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.051082 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.052528 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.054681 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.064817 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.065024 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.565003169 +0000 UTC m=+175.738316541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.065141 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.065696 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.565672755 +0000 UTC m=+175.738986287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.068357 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.166627 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.166829 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.666797279 +0000 UTC m=+175.840110641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.167137 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.167200 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.167266 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.167300 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7llc\" (UniqueName: \"kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.167649 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.667633669 +0000 UTC m=+175.840947051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.244107 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.245553 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.249001 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.257285 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.268019 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.268223 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.76819001 +0000 UTC m=+175.941503372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.268538 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.268668 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7llc\" (UniqueName: \"kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.268797 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.268873 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.268928 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.768917197 +0000 UTC m=+175.942230739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.269644 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.269655 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.298805 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7llc\" (UniqueName: \"kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc\") pod \"certified-operators-qw9xf\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.368872 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.369912 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.370187 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.370221 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.370260 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ql58\" (UniqueName: \"kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.370410 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.87038948 +0000 UTC m=+176.043702842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.407373 4727 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-k7cws container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.407918 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" podUID="422a3020-82d4-48e2-ae59-9ab06061c5ee" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.407375 4727 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-k7cws container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.408020 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" podUID="422a3020-82d4-48e2-ae59-9ab06061c5ee" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.456050 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.457278 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.471517 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.473412 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.473465 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.473516 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.473568 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ql58\" (UniqueName: \"kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.473734 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:05.973720536 +0000 UTC m=+176.147033898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.474239 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.474299 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.497438 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f5c419a2-0bf0-4646-9a7d-df769977d3ad","Type":"ContainerStarted","Data":"990e3389733b1004a1dc5f0ea56c7d1e061a6b7ab3d3262d87928f6ee8587a1e"} Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.509651 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ql58\" (UniqueName: \"kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58\") pod \"community-operators-zfscf\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.511300 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ht2kk" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.575294 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.577152 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.577581 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.077522343 +0000 UTC m=+176.250835705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.578021 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2c9m\" (UniqueName: \"kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.578243 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.578357 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.578596 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.583207 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.083177105 +0000 UTC m=+176.256490627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.656171 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.657253 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.683798 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.684570 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.684745 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.684832 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2c9m\" (UniqueName: \"kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.684867 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.685771 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.685848 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.185831616 +0000 UTC m=+176.359144978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.686045 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.708132 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2c9m\" (UniqueName: \"kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m\") pod \"certified-operators-98m5m\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.785822 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.786328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.786661 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.786690 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncbln\" (UniqueName: \"kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.787192 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.287177506 +0000 UTC m=+176.460490868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.805459 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.856534 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.857330 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.868516 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.868782 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.878499 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.887727 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.888099 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.888187 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.888254 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncbln\" (UniqueName: \"kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.888826 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.388805232 +0000 UTC m=+176.562118594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.889389 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.889669 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.926201 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.927288 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncbln\" (UniqueName: \"kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln\") pod \"community-operators-mm76q\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.988894 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:05 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:05 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:05 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.988960 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.992049 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.992207 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:05 crc kubenswrapper[4727]: I0929 10:25:05.992243 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:05 crc kubenswrapper[4727]: E0929 10:25:05.993568 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.493549341 +0000 UTC m=+176.666862703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.050686 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.075568 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.093846 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.094043 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.59400523 +0000 UTC m=+176.767318592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.094144 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.094231 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.094376 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.094855 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.095097 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.595074455 +0000 UTC m=+176.768388017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.129894 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.195477 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.195871 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.695831641 +0000 UTC m=+176.869145003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.231714 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.238782 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.298268 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.298708 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.798695176 +0000 UTC m=+176.972008538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.399418 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.399618 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.899579625 +0000 UTC m=+177.072892987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.399803 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.400168 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:06.900160489 +0000 UTC m=+177.073473851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.503387 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.503613 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.003581387 +0000 UTC m=+177.176894749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.503686 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.504471 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.004462928 +0000 UTC m=+177.177776290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.520603 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerStarted","Data":"a4bc973e3d99f21e67ef2d7af40f5187a47ec9ee492677cdf6904de986f1c183"} Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.544761 4727 generic.go:334] "Generic (PLEG): container finished" podID="f5c419a2-0bf0-4646-9a7d-df769977d3ad" containerID="990e3389733b1004a1dc5f0ea56c7d1e061a6b7ab3d3262d87928f6ee8587a1e" exitCode=0 Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.544851 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f5c419a2-0bf0-4646-9a7d-df769977d3ad","Type":"ContainerDied","Data":"990e3389733b1004a1dc5f0ea56c7d1e061a6b7ab3d3262d87928f6ee8587a1e"} Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.561784 4727 generic.go:334] "Generic (PLEG): container finished" podID="034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" containerID="ec3c7e62a378703afcf313eb626eaecf2011a5dcc0c5e2cbe9fc52b0bd80ecdc" exitCode=0 Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.561861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" event={"ID":"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2","Type":"ContainerDied","Data":"ec3c7e62a378703afcf313eb626eaecf2011a5dcc0c5e2cbe9fc52b0bd80ecdc"} Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.565723 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerStarted","Data":"1a216e208017aa434966bd9186189312ff935ff19f620f5fdc2eddba13b841f4"} Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.584749 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerStarted","Data":"ec3a2562d11ba5955b62aaad0a400e79b68f2811eab9503a9db4ef5d6e65c520"} Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.606224 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.606792 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.10676869 +0000 UTC m=+177.280082062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.707464 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.709221 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.209207535 +0000 UTC m=+177.382520897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.738418 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.809843 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.810382 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.31035796 +0000 UTC m=+177.483671322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.810645 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.811032 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.311025716 +0000 UTC m=+177.484339078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.871009 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.911583 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.911810 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.411778202 +0000 UTC m=+177.585091564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.911894 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:06 crc kubenswrapper[4727]: E0929 10:25:06.912226 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.412218992 +0000 UTC m=+177.585532354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.985995 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:06 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:06 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:06 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:06 crc kubenswrapper[4727]: I0929 10:25:06.986066 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.013057 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.013580 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.513555622 +0000 UTC m=+177.686868984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.115191 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.115622 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.615600508 +0000 UTC m=+177.788913870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.216974 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.217206 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.717171653 +0000 UTC m=+177.890485015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.217536 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.217893 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.717878569 +0000 UTC m=+177.891191931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.248329 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.249599 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.254897 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.268296 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.320303 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.320938 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.820899118 +0000 UTC m=+177.994212480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.321576 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vh4\" (UniqueName: \"kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.321664 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.321740 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.321817 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.322271 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.82226314 +0000 UTC m=+177.995576502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.423593 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.423822 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.923783814 +0000 UTC m=+178.097097186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.424005 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vh4\" (UniqueName: \"kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.424077 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.424114 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.424168 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.424619 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:07.924602853 +0000 UTC m=+178.097916215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.425561 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.425566 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.452547 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vh4\" (UniqueName: \"kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4\") pod \"redhat-marketplace-x887h\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.525956 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.526189 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.026129147 +0000 UTC m=+178.199442509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.527491 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.528051 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.028028631 +0000 UTC m=+178.201342163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.572966 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.602634 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17ddda0d-c994-4938-89c5-96b07e7cb218","Type":"ContainerStarted","Data":"82ba8435618049ee9ba77adaa521f001251b74808cc29e2e57192b6ffaabbef0"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.602718 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17ddda0d-c994-4938-89c5-96b07e7cb218","Type":"ContainerStarted","Data":"1653e703188245f44fb599bfadd60e7b221b311cdf304f0cc1f616d75ab24cf0"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.604843 4727 generic.go:334] "Generic (PLEG): container finished" podID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerID="b9ee8981ff63daf4348c55af8b3a9974643c6fd61e28deb96ce8e66345f57acf" exitCode=0 Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.605002 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerDied","Data":"b9ee8981ff63daf4348c55af8b3a9974643c6fd61e28deb96ce8e66345f57acf"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.607195 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.612779 4727 generic.go:334] "Generic (PLEG): container finished" podID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerID="2de2baca6fb82adccfcd528d514b646f7697bc0ad4a9d0663ca1d2588e09d559" exitCode=0 Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.614469 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerDied","Data":"2de2baca6fb82adccfcd528d514b646f7697bc0ad4a9d0663ca1d2588e09d559"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.629522 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.630051 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.130024586 +0000 UTC m=+178.303337948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.649807 4727 generic.go:334] "Generic (PLEG): container finished" podID="24076974-7055-499b-92f6-16ed1fd24121" containerID="9332668c92e54860a4354fcf3cac3de19118c73dfd051e966c6d8767f5930470" exitCode=0 Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.649938 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerDied","Data":"9332668c92e54860a4354fcf3cac3de19118c73dfd051e966c6d8767f5930470"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.651673 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.652886 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.668789 4727 generic.go:334] "Generic (PLEG): container finished" podID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerID="5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d" exitCode=0 Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.673307 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerDied","Data":"5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.673445 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerStarted","Data":"ee2580525b9c005d4d5914c0c78010cd3ba4774c6a86f76f284910d650742533"} Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.684630 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.731795 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8vpt\" (UniqueName: \"kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.731852 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.731899 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.732236 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.738447 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.238420021 +0000 UTC m=+178.411733563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.833001 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.833401 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8vpt\" (UniqueName: \"kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.833427 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.833449 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.833889 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.833973 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.333954634 +0000 UTC m=+178.507267996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.834446 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.852165 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8vpt\" (UniqueName: \"kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt\") pod \"redhat-marketplace-2tm27\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.935470 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:07 crc kubenswrapper[4727]: E0929 10:25:07.935849 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.435832737 +0000 UTC m=+178.609146099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.949320 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.978249 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.986270 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:07 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:07 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:07 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:07 crc kubenswrapper[4727]: I0929 10:25:07.986369 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.014078 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:25:08 crc kubenswrapper[4727]: W0929 10:25:08.028403 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03310bb6_f32c_48a0_a2df_a34b133b12c3.slice/crio-6bd12842e245b4c27d8910538a75cc9f59ec45013ec5daf91f742c3c13b32b9d WatchSource:0}: Error finding container 6bd12842e245b4c27d8910538a75cc9f59ec45013ec5daf91f742c3c13b32b9d: Status 404 returned error can't find the container with id 6bd12842e245b4c27d8910538a75cc9f59ec45013ec5daf91f742c3c13b32b9d Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.036499 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.037381 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.037479 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access\") pod \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.037531 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir\") pod \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\" (UID: \"f5c419a2-0bf0-4646-9a7d-df769977d3ad\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.038001 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f5c419a2-0bf0-4646-9a7d-df769977d3ad" (UID: "f5c419a2-0bf0-4646-9a7d-df769977d3ad"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.038112 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.538090088 +0000 UTC m=+178.711403450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.052917 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f5c419a2-0bf0-4646-9a7d-df769977d3ad" (UID: "f5c419a2-0bf0-4646-9a7d-df769977d3ad"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.139574 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume\") pod \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.140329 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fp9n\" (UniqueName: \"kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n\") pod \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.140388 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume\") pod \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\" (UID: \"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.141409 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" (UID: "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.150838 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.151062 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.151075 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.151085 4727 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5c419a2-0bf0-4646-9a7d-df769977d3ad-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.151442 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.651426267 +0000 UTC m=+178.824739629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.153439 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n" (OuterVolumeSpecName: "kube-api-access-7fp9n") pod "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" (UID: "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2"). InnerVolumeSpecName "kube-api-access-7fp9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.153495 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" (UID: "034c49df-ae6d-4483-b7cf-b5cd62bbb5a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.245134 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.252065 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.252565 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fp9n\" (UniqueName: \"kubernetes.io/projected/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-kube-api-access-7fp9n\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.252585 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.252668 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.752646594 +0000 UTC m=+178.925959956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.259179 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.259684 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c419a2-0bf0-4646-9a7d-df769977d3ad" containerName="pruner" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.259798 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c419a2-0bf0-4646-9a7d-df769977d3ad" containerName="pruner" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.259919 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" containerName="collect-profiles" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.260010 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" containerName="collect-profiles" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.260403 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5c419a2-0bf0-4646-9a7d-df769977d3ad" containerName="pruner" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.260557 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" containerName="collect-profiles" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.261985 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.270578 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.290666 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.359857 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.359920 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.359949 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8xl2\" (UniqueName: \"kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.359991 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.360350 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.860321701 +0000 UTC m=+179.033635053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.410880 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-k7cws" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.461253 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.461480 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.461561 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8xl2\" (UniqueName: \"kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.461621 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.461934 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:08.961899917 +0000 UTC m=+179.135213279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.462174 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.462553 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.479972 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8xl2\" (UniqueName: \"kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2\") pod \"redhat-operators-j7lzd\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.563680 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.564089 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.064070456 +0000 UTC m=+179.237383818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.590089 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.645653 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.647107 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.651073 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.665651 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.665886 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.165857516 +0000 UTC m=+179.339170868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.665976 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.666607 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.166599373 +0000 UTC m=+179.339912735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.686955 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f5c419a2-0bf0-4646-9a7d-df769977d3ad","Type":"ContainerDied","Data":"d800a459b0190ee67c5ebef557f6de972db383acf1ff7a1e5365ac22aecf2ffa"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.687296 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d800a459b0190ee67c5ebef557f6de972db383acf1ff7a1e5365ac22aecf2ffa" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.687517 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.689163 4727 generic.go:334] "Generic (PLEG): container finished" podID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerID="e32941b33f5599311c0b02c3c4b9646a9e3f4dd4bf87ebcea070ee203b3082c5" exitCode=0 Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.689216 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerDied","Data":"e32941b33f5599311c0b02c3c4b9646a9e3f4dd4bf87ebcea070ee203b3082c5"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.689237 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerStarted","Data":"6bd12842e245b4c27d8910538a75cc9f59ec45013ec5daf91f742c3c13b32b9d"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.691223 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" event={"ID":"034c49df-ae6d-4483-b7cf-b5cd62bbb5a2","Type":"ContainerDied","Data":"6fd452c1006309f0a895c3f2f487e80cc876b4a7016433f0a0bab2aaa285500c"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.691247 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fd452c1006309f0a895c3f2f487e80cc876b4a7016433f0a0bab2aaa285500c" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.691347 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.698267 4727 generic.go:334] "Generic (PLEG): container finished" podID="17ddda0d-c994-4938-89c5-96b07e7cb218" containerID="82ba8435618049ee9ba77adaa521f001251b74808cc29e2e57192b6ffaabbef0" exitCode=0 Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.698370 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17ddda0d-c994-4938-89c5-96b07e7cb218","Type":"ContainerDied","Data":"82ba8435618049ee9ba77adaa521f001251b74808cc29e2e57192b6ffaabbef0"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.701067 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerStarted","Data":"6174ed4a2da267155dfe8a25a6c6ab32d534a75fe39df8e7245149355904c564"} Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.767520 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.767680 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.267647316 +0000 UTC m=+179.440960678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.767751 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.767788 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gt59\" (UniqueName: \"kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.767877 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.767928 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.768420 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.268401203 +0000 UTC m=+179.441714565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.840636 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:25:08 crc kubenswrapper[4727]: W0929 10:25:08.852271 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ed07777_5c70_4338_b9a5_dc3f12aae019.slice/crio-b4f81a5137efc6b079e5526ed97415432b72bca1e06954b08f1c7e4d4e68c73c WatchSource:0}: Error finding container b4f81a5137efc6b079e5526ed97415432b72bca1e06954b08f1c7e4d4e68c73c: Status 404 returned error can't find the container with id b4f81a5137efc6b079e5526ed97415432b72bca1e06954b08f1c7e4d4e68c73c Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.869270 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.869423 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.369392555 +0000 UTC m=+179.542705917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.869622 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.869676 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.869737 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.869759 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gt59\" (UniqueName: \"kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.870367 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.870397 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.370369848 +0000 UTC m=+179.543683210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.870613 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.891066 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gt59\" (UniqueName: \"kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59\") pod \"redhat-operators-p52fg\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.971225 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.971420 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.47139868 +0000 UTC m=+179.644712042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.971976 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:08 crc kubenswrapper[4727]: E0929 10:25:08.972289 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.472280121 +0000 UTC m=+179.645593483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.972553 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.987704 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:08 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:08 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:08 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:08 crc kubenswrapper[4727]: I0929 10:25:08.987782 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.073973 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.074232 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.574195864 +0000 UTC m=+179.747509226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.074716 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.075302 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.575271679 +0000 UTC m=+179.748585061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.176482 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.176755 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.676715811 +0000 UTC m=+179.850029173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.176975 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.177382 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.677364806 +0000 UTC m=+179.850678168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.225984 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:25:09 crc kubenswrapper[4727]: W0929 10:25:09.228488 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeef57f61_5734_43ca_bd98_ba2e1131397e.slice/crio-da11f49f1a9a705d15b7f4598aaf41dc2d118c53decfa3c119a9e2bb6ace38cb WatchSource:0}: Error finding container da11f49f1a9a705d15b7f4598aaf41dc2d118c53decfa3c119a9e2bb6ace38cb: Status 404 returned error can't find the container with id da11f49f1a9a705d15b7f4598aaf41dc2d118c53decfa3c119a9e2bb6ace38cb Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.277985 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.278496 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.778127202 +0000 UTC m=+179.951440564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.278800 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.279376 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.779359531 +0000 UTC m=+179.952672893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.380057 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.380552 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.880503706 +0000 UTC m=+180.053817068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.380942 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.381370 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.881331775 +0000 UTC m=+180.054645137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.482412 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.482619 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.982589493 +0000 UTC m=+180.155902855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.482953 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.483412 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:09.983393272 +0000 UTC m=+180.156706814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.518052 4727 patch_prober.go:28] interesting pod/console-f9d7485db-fkfsn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.518111 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fkfsn" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.584018 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.584410 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.084361003 +0000 UTC m=+180.257674405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.584734 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.585147 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.085128951 +0000 UTC m=+180.258442503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.684707 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.685747 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.685871 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.185851426 +0000 UTC m=+180.359164788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.686025 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.686310 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.186303926 +0000 UTC m=+180.359617288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.712614 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerStarted","Data":"da11f49f1a9a705d15b7f4598aaf41dc2d118c53decfa3c119a9e2bb6ace38cb"} Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.713934 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerStarted","Data":"b4f81a5137efc6b079e5526ed97415432b72bca1e06954b08f1c7e4d4e68c73c"} Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.716151 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerID="bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2" exitCode=0 Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.716233 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerDied","Data":"bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2"} Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.718885 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" event={"ID":"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee","Type":"ContainerStarted","Data":"fd6bb568676f2755ee0bacf07fe16dc2692c07a4c147bd7bb41dae6b5b8feefc"} Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.787224 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.787614 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.287575634 +0000 UTC m=+180.460889006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.787866 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.788245 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.28823626 +0000 UTC m=+180.461549622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.888850 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.889035 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.389000066 +0000 UTC m=+180.562313448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.889196 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.889542 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.389531608 +0000 UTC m=+180.562844980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.970035 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.986740 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:09 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:09 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:09 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.987082 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.994900 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.995200 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.495153918 +0000 UTC m=+180.668467300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.995598 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:09 crc kubenswrapper[4727]: E0929 10:25:09.996113 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.49610193 +0000 UTC m=+180.669415302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:09 crc kubenswrapper[4727]: I0929 10:25:09.998987 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.016382 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vvk6t" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.096881 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access\") pod \"17ddda0d-c994-4938-89c5-96b07e7cb218\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.097260 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.097392 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir\") pod \"17ddda0d-c994-4938-89c5-96b07e7cb218\" (UID: \"17ddda0d-c994-4938-89c5-96b07e7cb218\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.097456 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.59743386 +0000 UTC m=+180.770747222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.097524 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "17ddda0d-c994-4938-89c5-96b07e7cb218" (UID: "17ddda0d-c994-4938-89c5-96b07e7cb218"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.097880 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.098182 4727 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17ddda0d-c994-4938-89c5-96b07e7cb218-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.098702 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.598678189 +0000 UTC m=+180.771991551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.107573 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "17ddda0d-c994-4938-89c5-96b07e7cb218" (UID: "17ddda0d-c994-4938-89c5-96b07e7cb218"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.200171 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.200468 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.700426178 +0000 UTC m=+180.873739540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.200601 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.200690 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17ddda0d-c994-4938-89c5-96b07e7cb218-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.201090 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.701077303 +0000 UTC m=+180.874390865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.245968 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.253026 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ql82q" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.301608 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.301827 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.801781858 +0000 UTC m=+180.975095220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.302156 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.302557 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.802541316 +0000 UTC m=+180.975854678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.403464 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.404405 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:10.904382487 +0000 UTC m=+181.077695859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.439544 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.439609 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.439669 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.439732 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.505052 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.505934 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.005917971 +0000 UTC m=+181.179231333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.574635 4727 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hr9hg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]log ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]etcd ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/generic-apiserver-start-informers ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/max-in-flight-filter ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/image.openshift.io-apiserver-caches ok Sep 29 10:25:10 crc kubenswrapper[4727]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Sep 29 10:25:10 crc kubenswrapper[4727]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/project.openshift.io-projectcache ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/openshift.io-startinformers ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/openshift.io-restmapperupdater ok Sep 29 10:25:10 crc kubenswrapper[4727]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 29 10:25:10 crc kubenswrapper[4727]: livez check failed Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.574728 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" podUID="04eb68eb-eab0-4868-96e6-38781e313d3a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.606698 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.607218 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.107193679 +0000 UTC m=+181.280507041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.609766 4727 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.645178 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x7jzw" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.709070 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.709531 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.209513912 +0000 UTC m=+181.382827474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.725006 4727 generic.go:334] "Generic (PLEG): container finished" podID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerID="5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8" exitCode=0 Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.725084 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerDied","Data":"5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8"} Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.728128 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17ddda0d-c994-4938-89c5-96b07e7cb218","Type":"ContainerDied","Data":"1653e703188245f44fb599bfadd60e7b221b311cdf304f0cc1f616d75ab24cf0"} Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.728180 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1653e703188245f44fb599bfadd60e7b221b311cdf304f0cc1f616d75ab24cf0" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.728251 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.731129 4727 generic.go:334] "Generic (PLEG): container finished" podID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerID="0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e" exitCode=0 Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.731216 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerDied","Data":"0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e"} Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.735463 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" event={"ID":"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee","Type":"ContainerStarted","Data":"83bedc97cc8b7393c469116cfa037497e10df8b56ca8bbd49719f7f369193b79"} Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.810150 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.810429 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.310393911 +0000 UTC m=+181.483707273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.810861 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.811635 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.311625789 +0000 UTC m=+181.484939151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.911680 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.911810 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.411788761 +0000 UTC m=+181.585102123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.912225 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:10 crc kubenswrapper[4727]: E0929 10:25:10.912796 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.412771584 +0000 UTC m=+181.586084946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.986237 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:10 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:10 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:10 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:10 crc kubenswrapper[4727]: I0929 10:25:10.986304 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.013303 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:11 crc kubenswrapper[4727]: E0929 10:25:11.013763 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.513743455 +0000 UTC m=+181.687056817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.114989 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: E0929 10:25:11.115446 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.615425853 +0000 UTC m=+181.788739215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.216609 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:11 crc kubenswrapper[4727]: E0929 10:25:11.216767 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.716740012 +0000 UTC m=+181.890053374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.216878 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: E0929 10:25:11.217362 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.717317686 +0000 UTC m=+181.890631238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nvqqb" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.318565 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:11 crc kubenswrapper[4727]: E0929 10:25:11.319199 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 10:25:11.819164377 +0000 UTC m=+181.992477739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.333455 4727 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-29T10:25:10.60979207Z","Handler":null,"Name":""} Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.338206 4727 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.338252 4727 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.397253 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-lljzg" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.423358 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.431024 4727 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.431073 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.487385 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nvqqb\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.525047 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.529244 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.588481 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.597216 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.758436 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" event={"ID":"7538e768-8b1c-4fd3-88b4-a2ab8fdbd3ee","Type":"ContainerStarted","Data":"3abc3709a4ff60ac79d0900331599cafb08444ebeb559e0e344620492753a609"} Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.781374 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9mtzk" podStartSLOduration=24.781303392 podStartE2EDuration="24.781303392s" podCreationTimestamp="2025-09-29 10:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:11.777184056 +0000 UTC m=+181.950497428" watchObservedRunningTime="2025-09-29 10:25:11.781303392 +0000 UTC m=+181.954616754" Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.853793 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.986564 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:11 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:11 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:11 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:11 crc kubenswrapper[4727]: I0929 10:25:11.986662 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:12 crc kubenswrapper[4727]: I0929 10:25:12.766779 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" event={"ID":"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0","Type":"ContainerStarted","Data":"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9"} Sep 29 10:25:12 crc kubenswrapper[4727]: I0929 10:25:12.767415 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:12 crc kubenswrapper[4727]: I0929 10:25:12.767437 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" event={"ID":"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0","Type":"ContainerStarted","Data":"990ddb3b196207b633cbb2901f3d7842d3d2a45e20538da56fcb3c5c40333519"} Sep 29 10:25:12 crc kubenswrapper[4727]: I0929 10:25:12.985974 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:12 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:12 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:12 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:12 crc kubenswrapper[4727]: I0929 10:25:12.986060 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:13 crc kubenswrapper[4727]: I0929 10:25:13.118114 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 29 10:25:13 crc kubenswrapper[4727]: I0929 10:25:13.984961 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:13 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:13 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:13 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:13 crc kubenswrapper[4727]: I0929 10:25:13.985089 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:14 crc kubenswrapper[4727]: I0929 10:25:14.639215 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:25:14 crc kubenswrapper[4727]: I0929 10:25:14.643864 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hr9hg" Sep 29 10:25:14 crc kubenswrapper[4727]: I0929 10:25:14.661856 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" podStartSLOduration=157.661835366 podStartE2EDuration="2m37.661835366s" podCreationTimestamp="2025-09-29 10:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:25:12.789064786 +0000 UTC m=+182.962378148" watchObservedRunningTime="2025-09-29 10:25:14.661835366 +0000 UTC m=+184.835148728" Sep 29 10:25:14 crc kubenswrapper[4727]: I0929 10:25:14.984259 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:14 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:14 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:14 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:14 crc kubenswrapper[4727]: I0929 10:25:14.984701 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:15 crc kubenswrapper[4727]: I0929 10:25:15.984839 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:15 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:15 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:15 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:15 crc kubenswrapper[4727]: I0929 10:25:15.984927 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:16 crc kubenswrapper[4727]: I0929 10:25:16.985522 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:16 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:16 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:16 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:16 crc kubenswrapper[4727]: I0929 10:25:16.985630 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:17 crc kubenswrapper[4727]: I0929 10:25:17.984059 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:17 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:17 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:17 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:17 crc kubenswrapper[4727]: I0929 10:25:17.984131 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:18 crc kubenswrapper[4727]: I0929 10:25:18.985272 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:18 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:18 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:18 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:18 crc kubenswrapper[4727]: I0929 10:25:18.985345 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:19 crc kubenswrapper[4727]: I0929 10:25:19.246389 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:25:19 crc kubenswrapper[4727]: I0929 10:25:19.246781 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:25:19 crc kubenswrapper[4727]: I0929 10:25:19.517719 4727 patch_prober.go:28] interesting pod/console-f9d7485db-fkfsn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Sep 29 10:25:19 crc kubenswrapper[4727]: I0929 10:25:19.517786 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fkfsn" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.070586 4727 patch_prober.go:28] interesting pod/router-default-5444994796-t4lhc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 10:25:20 crc kubenswrapper[4727]: [-]has-synced failed: reason withheld Sep 29 10:25:20 crc kubenswrapper[4727]: [+]process-running ok Sep 29 10:25:20 crc kubenswrapper[4727]: healthz check failed Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.070680 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t4lhc" podUID="274d4012-934c-426c-8902-20d19732cfea" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.439174 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.439237 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.439438 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.439552 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.985811 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:25:20 crc kubenswrapper[4727]: I0929 10:25:20.987724 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-t4lhc" Sep 29 10:25:29 crc kubenswrapper[4727]: I0929 10:25:29.522955 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:25:29 crc kubenswrapper[4727]: I0929 10:25:29.527376 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.439095 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.439487 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.439188 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.439544 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.439565 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.440742 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.440798 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.440922 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"7f2150e57aec7ee75945821e9672c7989096aa20803a8ed93edf792ad10cc252"} pod="openshift-console/downloads-7954f5f757-2qq2r" containerMessage="Container download-server failed liveness probe, will be restarted" Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.441217 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" containerID="cri-o://7f2150e57aec7ee75945821e9672c7989096aa20803a8ed93edf792ad10cc252" gracePeriod=2 Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.892182 4727 generic.go:334] "Generic (PLEG): container finished" podID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerID="7f2150e57aec7ee75945821e9672c7989096aa20803a8ed93edf792ad10cc252" exitCode=0 Sep 29 10:25:30 crc kubenswrapper[4727]: I0929 10:25:30.892234 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2qq2r" event={"ID":"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451","Type":"ContainerDied","Data":"7f2150e57aec7ee75945821e9672c7989096aa20803a8ed93edf792ad10cc252"} Sep 29 10:25:31 crc kubenswrapper[4727]: I0929 10:25:31.603440 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:25:40 crc kubenswrapper[4727]: I0929 10:25:40.242319 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fzb2k" Sep 29 10:25:40 crc kubenswrapper[4727]: I0929 10:25:40.439860 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:40 crc kubenswrapper[4727]: I0929 10:25:40.439940 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:49 crc kubenswrapper[4727]: I0929 10:25:49.246965 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:25:49 crc kubenswrapper[4727]: I0929 10:25:49.248682 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:25:49 crc kubenswrapper[4727]: I0929 10:25:49.248806 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:25:49 crc kubenswrapper[4727]: I0929 10:25:49.249606 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:25:49 crc kubenswrapper[4727]: I0929 10:25:49.249667 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47" gracePeriod=600 Sep 29 10:25:50 crc kubenswrapper[4727]: I0929 10:25:50.438901 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:25:50 crc kubenswrapper[4727]: I0929 10:25:50.438984 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:25:55 crc kubenswrapper[4727]: E0929 10:25:55.454217 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 29 10:25:55 crc kubenswrapper[4727]: E0929 10:25:55.455023 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ql58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zfscf_openshift-marketplace(d2dcef0d-c089-4c5a-967b-19beeaceb1de): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:25:55 crc kubenswrapper[4727]: E0929 10:25:55.456554 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zfscf" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" Sep 29 10:26:00 crc kubenswrapper[4727]: I0929 10:26:00.441736 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:00 crc kubenswrapper[4727]: I0929 10:26:00.442411 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:02 crc kubenswrapper[4727]: E0929 10:26:02.163612 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zfscf" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" Sep 29 10:26:03 crc kubenswrapper[4727]: I0929 10:26:03.103485 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47" exitCode=0 Sep 29 10:26:03 crc kubenswrapper[4727]: I0929 10:26:03.103582 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47"} Sep 29 10:26:07 crc kubenswrapper[4727]: E0929 10:26:07.996395 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 29 10:26:07 crc kubenswrapper[4727]: E0929 10:26:07.996940 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x7llc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qw9xf_openshift-marketplace(24076974-7055-499b-92f6-16ed1fd24121): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:07 crc kubenswrapper[4727]: E0929 10:26:07.998193 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qw9xf" podUID="24076974-7055-499b-92f6-16ed1fd24121" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.521594 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qw9xf" podUID="24076974-7055-499b-92f6-16ed1fd24121" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.569821 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.570027 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8vpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2tm27_openshift-marketplace(cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.571636 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2tm27" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.595908 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.596125 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ncbln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mm76q_openshift-marketplace(b5b4ca6c-b281-45ab-9417-d2903007a4f8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:08 crc kubenswrapper[4727]: E0929 10:26:08.597394 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mm76q" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" Sep 29 10:26:10 crc kubenswrapper[4727]: I0929 10:26:10.440107 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:10 crc kubenswrapper[4727]: I0929 10:26:10.440193 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.535585 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mm76q" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.569744 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.569945 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8xl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-j7lzd_openshift-marketplace(1ed07777-5c70-4338-b9a5-dc3f12aae019): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.571227 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-j7lzd" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.619493 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.619657 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8gt59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-p52fg_openshift-marketplace(eef57f61-5734-43ca-bd98-ba2e1131397e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.621032 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-p52fg" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.629719 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.629877 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t2c9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-98m5m_openshift-marketplace(66fb25b1-5331-4450-87a0-9e032e6ef984): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.631364 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-98m5m" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.664377 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.664634 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4vh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-x887h_openshift-marketplace(03310bb6-f32c-48a0-a2df-a34b133b12c3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:26:11 crc kubenswrapper[4727]: E0929 10:26:11.666016 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-x887h" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" Sep 29 10:26:12 crc kubenswrapper[4727]: I0929 10:26:12.157830 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2qq2r" event={"ID":"7f67f7b2-5a7d-421a-aa4a-333ef6d1e451","Type":"ContainerStarted","Data":"ac0a1d61edbd69a98fdf0bfec70fee5277d36ea935c48b4c3dec3bff831baac7"} Sep 29 10:26:12 crc kubenswrapper[4727]: I0929 10:26:12.158400 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:26:12 crc kubenswrapper[4727]: I0929 10:26:12.158528 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:12 crc kubenswrapper[4727]: I0929 10:26:12.158617 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:12 crc kubenswrapper[4727]: I0929 10:26:12.160631 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc"} Sep 29 10:26:12 crc kubenswrapper[4727]: E0929 10:26:12.163750 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-98m5m" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" Sep 29 10:26:12 crc kubenswrapper[4727]: E0929 10:26:12.163977 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-j7lzd" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" Sep 29 10:26:12 crc kubenswrapper[4727]: E0929 10:26:12.164056 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-p52fg" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" Sep 29 10:26:13 crc kubenswrapper[4727]: I0929 10:26:13.167210 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:13 crc kubenswrapper[4727]: I0929 10:26:13.167281 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:16 crc kubenswrapper[4727]: I0929 10:26:16.187377 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerStarted","Data":"0d306692a506ead0120b9a11a69496374b269a0d692c90515ccfc05ef6c59f43"} Sep 29 10:26:17 crc kubenswrapper[4727]: I0929 10:26:17.196309 4727 generic.go:334] "Generic (PLEG): container finished" podID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerID="0d306692a506ead0120b9a11a69496374b269a0d692c90515ccfc05ef6c59f43" exitCode=0 Sep 29 10:26:17 crc kubenswrapper[4727]: I0929 10:26:17.196402 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerDied","Data":"0d306692a506ead0120b9a11a69496374b269a0d692c90515ccfc05ef6c59f43"} Sep 29 10:26:20 crc kubenswrapper[4727]: I0929 10:26:20.439222 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:20 crc kubenswrapper[4727]: I0929 10:26:20.439693 4727 patch_prober.go:28] interesting pod/downloads-7954f5f757-2qq2r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Sep 29 10:26:20 crc kubenswrapper[4727]: I0929 10:26:20.439731 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:20 crc kubenswrapper[4727]: I0929 10:26:20.439781 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2qq2r" podUID="7f67f7b2-5a7d-421a-aa4a-333ef6d1e451" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Sep 29 10:26:29 crc kubenswrapper[4727]: I0929 10:26:29.265874 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerStarted","Data":"579a705e128d65b62e9872cf42356e3bceb81b18be7e17938ba3297680c15dda"} Sep 29 10:26:30 crc kubenswrapper[4727]: I0929 10:26:30.472895 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2qq2r" Sep 29 10:26:31 crc kubenswrapper[4727]: I0929 10:26:31.295259 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfscf" podStartSLOduration=6.425542818 podStartE2EDuration="1m26.295241018s" podCreationTimestamp="2025-09-29 10:25:05 +0000 UTC" firstStartedPulling="2025-09-29 10:25:07.623922893 +0000 UTC m=+177.797236255" lastFinishedPulling="2025-09-29 10:26:27.493621093 +0000 UTC m=+257.666934455" observedRunningTime="2025-09-29 10:26:31.29102469 +0000 UTC m=+261.464338052" watchObservedRunningTime="2025-09-29 10:26:31.295241018 +0000 UTC m=+261.468554380" Sep 29 10:26:35 crc kubenswrapper[4727]: I0929 10:26:35.576185 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:26:35 crc kubenswrapper[4727]: I0929 10:26:35.577157 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:26:37 crc kubenswrapper[4727]: I0929 10:26:37.320580 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:26:37 crc kubenswrapper[4727]: I0929 10:26:37.398278 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.068427 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.072126 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.155674 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.170421 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.170482 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.170513 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.173745 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.174992 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.183419 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.192418 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.196188 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.197393 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.429206 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.440787 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:26:47 crc kubenswrapper[4727]: I0929 10:26:47.480489 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 10:27:13 crc kubenswrapper[4727]: W0929 10:27:13.649671 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-57cf0502c2ce2efd6e5d86d04296e18c68f1210ffdd74c507acae9ddfbd94adc WatchSource:0}: Error finding container 57cf0502c2ce2efd6e5d86d04296e18c68f1210ffdd74c507acae9ddfbd94adc: Status 404 returned error can't find the container with id 57cf0502c2ce2efd6e5d86d04296e18c68f1210ffdd74c507acae9ddfbd94adc Sep 29 10:27:13 crc kubenswrapper[4727]: W0929 10:27:13.960785 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b987fc6da73ec3c616f3e0e1f31fc8479aa5e7fe4a95d8582c642f4b40a5ef57 WatchSource:0}: Error finding container b987fc6da73ec3c616f3e0e1f31fc8479aa5e7fe4a95d8582c642f4b40a5ef57: Status 404 returned error can't find the container with id b987fc6da73ec3c616f3e0e1f31fc8479aa5e7fe4a95d8582c642f4b40a5ef57 Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.572861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b987fc6da73ec3c616f3e0e1f31fc8479aa5e7fe4a95d8582c642f4b40a5ef57"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.575029 4727 generic.go:334] "Generic (PLEG): container finished" podID="24076974-7055-499b-92f6-16ed1fd24121" containerID="d830a17a298fc873f4cae63e5ba32484de9d8855768f9eb12a82bc5ab82d9eea" exitCode=0 Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.575172 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerDied","Data":"d830a17a298fc873f4cae63e5ba32484de9d8855768f9eb12a82bc5ab82d9eea"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.585030 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"21abfd0b120f2460df87f28805cc3f0bbe05647ea9767ebcd587e3f23abfc575"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.585086 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c77ab1032f5cef1001a464b1f9a0d327b757e62682e455bd805e1050e5ac2c61"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.587709 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerStarted","Data":"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.604709 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"63b1b93c85d4708c8734b69bd9992076882da7bd514aadf94b3c71316747c8da"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.604781 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"57cf0502c2ce2efd6e5d86d04296e18c68f1210ffdd74c507acae9ddfbd94adc"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.606975 4727 generic.go:334] "Generic (PLEG): container finished" podID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerID="1a59db951f74e7ba6e61b3146ef587284a21b970641c30fc4463aa5c554463cf" exitCode=0 Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.607067 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerDied","Data":"1a59db951f74e7ba6e61b3146ef587284a21b970641c30fc4463aa5c554463cf"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.612360 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerStarted","Data":"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.617742 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerStarted","Data":"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.620859 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerID="7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131" exitCode=0 Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.620953 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerDied","Data":"7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131"} Sep 29 10:27:14 crc kubenswrapper[4727]: I0929 10:27:14.625375 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerStarted","Data":"303f982ce9428d3b6d6a76946d5a632b45454e1186bbb7f2f1d5795ba4f7816f"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.634537 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerStarted","Data":"306ea6f8be6205841cb9a3ea5c467a82888c078ece596c702ff20cc79b9201b9"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.639122 4727 generic.go:334] "Generic (PLEG): container finished" podID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerID="170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea" exitCode=0 Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.639661 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerDied","Data":"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.644760 4727 generic.go:334] "Generic (PLEG): container finished" podID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerID="3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76" exitCode=0 Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.644866 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerDied","Data":"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.649850 4727 generic.go:334] "Generic (PLEG): container finished" podID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerID="4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38" exitCode=0 Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.651157 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerDied","Data":"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.662145 4727 generic.go:334] "Generic (PLEG): container finished" podID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerID="303f982ce9428d3b6d6a76946d5a632b45454e1186bbb7f2f1d5795ba4f7816f" exitCode=0 Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.662303 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerDied","Data":"303f982ce9428d3b6d6a76946d5a632b45454e1186bbb7f2f1d5795ba4f7816f"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.668428 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerStarted","Data":"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.675393 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"64ec12fcfaf5d94cfd1a7c2a3b47fb79f6ffda5cba67c99ee30ba47f7b5ed14c"} Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.675594 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:27:15 crc kubenswrapper[4727]: I0929 10:27:15.686328 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qw9xf" podStartSLOduration=3.108569508 podStartE2EDuration="2m10.6863032s" podCreationTimestamp="2025-09-29 10:25:05 +0000 UTC" firstStartedPulling="2025-09-29 10:25:07.656824693 +0000 UTC m=+177.830138055" lastFinishedPulling="2025-09-29 10:27:15.234558365 +0000 UTC m=+305.407871747" observedRunningTime="2025-09-29 10:27:15.661315498 +0000 UTC m=+305.834628870" watchObservedRunningTime="2025-09-29 10:27:15.6863032 +0000 UTC m=+305.859616562" Sep 29 10:27:16 crc kubenswrapper[4727]: I0929 10:27:16.696595 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerStarted","Data":"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7"} Sep 29 10:27:16 crc kubenswrapper[4727]: I0929 10:27:16.734225 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2tm27" podStartSLOduration=5.154823277 podStartE2EDuration="2m9.734197845s" podCreationTimestamp="2025-09-29 10:25:07 +0000 UTC" firstStartedPulling="2025-09-29 10:25:10.7402301 +0000 UTC m=+180.913543462" lastFinishedPulling="2025-09-29 10:27:15.319604678 +0000 UTC m=+305.492918030" observedRunningTime="2025-09-29 10:27:15.870680252 +0000 UTC m=+306.043993634" watchObservedRunningTime="2025-09-29 10:27:16.734197845 +0000 UTC m=+306.907511207" Sep 29 10:27:16 crc kubenswrapper[4727]: I0929 10:27:16.737281 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p52fg" podStartSLOduration=3.21787384 podStartE2EDuration="2m8.737262364s" podCreationTimestamp="2025-09-29 10:25:08 +0000 UTC" firstStartedPulling="2025-09-29 10:25:10.726902168 +0000 UTC m=+180.900215540" lastFinishedPulling="2025-09-29 10:27:16.246290702 +0000 UTC m=+306.419604064" observedRunningTime="2025-09-29 10:27:16.731981658 +0000 UTC m=+306.905295020" watchObservedRunningTime="2025-09-29 10:27:16.737262364 +0000 UTC m=+306.910575736" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.703847 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerStarted","Data":"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78"} Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.706366 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerStarted","Data":"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3"} Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.711685 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerStarted","Data":"f4d1f8e6cc859b5c6cf1fa37d852f164ba166f4cfed768742f721e63f47cc76d"} Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.714940 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerStarted","Data":"774fa49255e95b0556fd6c1381fc2fedc9a298242cbfce4d782b6041e36e3cd4"} Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.730915 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mm76q" podStartSLOduration=3.765710275 podStartE2EDuration="2m12.730889025s" podCreationTimestamp="2025-09-29 10:25:05 +0000 UTC" firstStartedPulling="2025-09-29 10:25:07.681901009 +0000 UTC m=+177.855214371" lastFinishedPulling="2025-09-29 10:27:16.647079759 +0000 UTC m=+306.820393121" observedRunningTime="2025-09-29 10:27:17.72756813 +0000 UTC m=+307.900881512" watchObservedRunningTime="2025-09-29 10:27:17.730889025 +0000 UTC m=+307.904202387" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.755740 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-98m5m" podStartSLOduration=3.729872827 podStartE2EDuration="2m12.755714982s" podCreationTimestamp="2025-09-29 10:25:05 +0000 UTC" firstStartedPulling="2025-09-29 10:25:07.606959897 +0000 UTC m=+177.780273259" lastFinishedPulling="2025-09-29 10:27:16.632802052 +0000 UTC m=+306.806115414" observedRunningTime="2025-09-29 10:27:17.750857478 +0000 UTC m=+307.924170850" watchObservedRunningTime="2025-09-29 10:27:17.755714982 +0000 UTC m=+307.929028344" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.797196 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x887h" podStartSLOduration=4.028693859 podStartE2EDuration="2m10.797167446s" podCreationTimestamp="2025-09-29 10:25:07 +0000 UTC" firstStartedPulling="2025-09-29 10:25:09.721436568 +0000 UTC m=+179.894749970" lastFinishedPulling="2025-09-29 10:27:16.489910195 +0000 UTC m=+306.663223557" observedRunningTime="2025-09-29 10:27:17.79652296 +0000 UTC m=+307.969836342" watchObservedRunningTime="2025-09-29 10:27:17.797167446 +0000 UTC m=+307.970480808" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.803275 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j7lzd" podStartSLOduration=3.986423129 podStartE2EDuration="2m9.803246812s" podCreationTimestamp="2025-09-29 10:25:08 +0000 UTC" firstStartedPulling="2025-09-29 10:25:10.733689787 +0000 UTC m=+180.907003149" lastFinishedPulling="2025-09-29 10:27:16.55051347 +0000 UTC m=+306.723826832" observedRunningTime="2025-09-29 10:27:17.776062495 +0000 UTC m=+307.949375857" watchObservedRunningTime="2025-09-29 10:27:17.803246812 +0000 UTC m=+307.976560174" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.980462 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:17 crc kubenswrapper[4727]: I0929 10:27:17.980846 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:18 crc kubenswrapper[4727]: I0929 10:27:18.030415 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:18 crc kubenswrapper[4727]: I0929 10:27:18.592037 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:27:18 crc kubenswrapper[4727]: I0929 10:27:18.592393 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:27:18 crc kubenswrapper[4727]: I0929 10:27:18.973461 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:18 crc kubenswrapper[4727]: I0929 10:27:18.974028 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:19 crc kubenswrapper[4727]: I0929 10:27:19.647703 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j7lzd" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="registry-server" probeResult="failure" output=< Sep 29 10:27:19 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 10:27:19 crc kubenswrapper[4727]: > Sep 29 10:27:20 crc kubenswrapper[4727]: I0929 10:27:20.016729 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p52fg" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="registry-server" probeResult="failure" output=< Sep 29 10:27:20 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 10:27:20 crc kubenswrapper[4727]: > Sep 29 10:27:20 crc kubenswrapper[4727]: I0929 10:27:20.935818 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.369845 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.370286 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.411829 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.804089 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.806364 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.806402 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:25 crc kubenswrapper[4727]: I0929 10:27:25.854125 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:26 crc kubenswrapper[4727]: I0929 10:27:26.052226 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:26 crc kubenswrapper[4727]: I0929 10:27:26.052276 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:26 crc kubenswrapper[4727]: I0929 10:27:26.092107 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:26 crc kubenswrapper[4727]: I0929 10:27:26.809464 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:26 crc kubenswrapper[4727]: I0929 10:27:26.812227 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:27 crc kubenswrapper[4727]: I0929 10:27:27.408154 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:27:27 crc kubenswrapper[4727]: I0929 10:27:27.573647 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:27:27 crc kubenswrapper[4727]: I0929 10:27:27.573744 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:27:27 crc kubenswrapper[4727]: I0929 10:27:27.614175 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:27:27 crc kubenswrapper[4727]: I0929 10:27:27.816126 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.019324 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.407560 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.628149 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.702640 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.785051 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mm76q" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="registry-server" containerID="cri-o://c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78" gracePeriod=2 Sep 29 10:27:28 crc kubenswrapper[4727]: I0929 10:27:28.785746 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-98m5m" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="registry-server" containerID="cri-o://f4d1f8e6cc859b5c6cf1fa37d852f164ba166f4cfed768742f721e63f47cc76d" gracePeriod=2 Sep 29 10:27:29 crc kubenswrapper[4727]: I0929 10:27:29.015297 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:29 crc kubenswrapper[4727]: I0929 10:27:29.058648 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:30 crc kubenswrapper[4727]: I0929 10:27:30.798187 4727 generic.go:334] "Generic (PLEG): container finished" podID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerID="f4d1f8e6cc859b5c6cf1fa37d852f164ba166f4cfed768742f721e63f47cc76d" exitCode=0 Sep 29 10:27:30 crc kubenswrapper[4727]: I0929 10:27:30.798250 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerDied","Data":"f4d1f8e6cc859b5c6cf1fa37d852f164ba166f4cfed768742f721e63f47cc76d"} Sep 29 10:27:30 crc kubenswrapper[4727]: I0929 10:27:30.807591 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:27:30 crc kubenswrapper[4727]: I0929 10:27:30.807875 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2tm27" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="registry-server" containerID="cri-o://de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d" gracePeriod=2 Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.727504 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.732910 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.803202 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.811954 4727 generic.go:334] "Generic (PLEG): container finished" podID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerID="c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78" exitCode=0 Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.812013 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerDied","Data":"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78"} Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.812045 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mm76q" event={"ID":"b5b4ca6c-b281-45ab-9417-d2903007a4f8","Type":"ContainerDied","Data":"ee2580525b9c005d4d5914c0c78010cd3ba4774c6a86f76f284910d650742533"} Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.812067 4727 scope.go:117] "RemoveContainer" containerID="c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.812196 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mm76q" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.822233 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98m5m" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.822553 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98m5m" event={"ID":"66fb25b1-5331-4450-87a0-9e032e6ef984","Type":"ContainerDied","Data":"1a216e208017aa434966bd9186189312ff935ff19f620f5fdc2eddba13b841f4"} Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.827258 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerID="de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d" exitCode=0 Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.827314 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerDied","Data":"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d"} Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.827368 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tm27" event={"ID":"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c","Type":"ContainerDied","Data":"6174ed4a2da267155dfe8a25a6c6ab32d534a75fe39df8e7245149355904c564"} Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.827460 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tm27" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.835892 4727 scope.go:117] "RemoveContainer" containerID="170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.851237 4727 scope.go:117] "RemoveContainer" containerID="5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861566 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2c9m\" (UniqueName: \"kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m\") pod \"66fb25b1-5331-4450-87a0-9e032e6ef984\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861696 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities\") pod \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861815 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content\") pod \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861849 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncbln\" (UniqueName: \"kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln\") pod \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861893 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities\") pod \"66fb25b1-5331-4450-87a0-9e032e6ef984\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861916 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content\") pod \"66fb25b1-5331-4450-87a0-9e032e6ef984\" (UID: \"66fb25b1-5331-4450-87a0-9e032e6ef984\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.861983 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities\") pod \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.862039 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content\") pod \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\" (UID: \"b5b4ca6c-b281-45ab-9417-d2903007a4f8\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.862078 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8vpt\" (UniqueName: \"kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt\") pod \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\" (UID: \"cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c\") " Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.862666 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities" (OuterVolumeSpecName: "utilities") pod "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" (UID: "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.863216 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities" (OuterVolumeSpecName: "utilities") pod "66fb25b1-5331-4450-87a0-9e032e6ef984" (UID: "66fb25b1-5331-4450-87a0-9e032e6ef984"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.863229 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities" (OuterVolumeSpecName: "utilities") pod "b5b4ca6c-b281-45ab-9417-d2903007a4f8" (UID: "b5b4ca6c-b281-45ab-9417-d2903007a4f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.866120 4727 scope.go:117] "RemoveContainer" containerID="c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.866949 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78\": container with ID starting with c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78 not found: ID does not exist" containerID="c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.866996 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78"} err="failed to get container status \"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78\": rpc error: code = NotFound desc = could not find container \"c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78\": container with ID starting with c54a6e4e5d462bea681c80a6db06309481e788ec59aaad95148d2e17f2fccb78 not found: ID does not exist" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.867034 4727 scope.go:117] "RemoveContainer" containerID="170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.867373 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea\": container with ID starting with 170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea not found: ID does not exist" containerID="170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.867395 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea"} err="failed to get container status \"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea\": rpc error: code = NotFound desc = could not find container \"170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea\": container with ID starting with 170b83b51f71be18b1d664b689ff015e88a7da736a80d0156a4a0ab3a2b7f2ea not found: ID does not exist" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.867408 4727 scope.go:117] "RemoveContainer" containerID="5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.867634 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d\": container with ID starting with 5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d not found: ID does not exist" containerID="5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.867657 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d"} err="failed to get container status \"5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d\": rpc error: code = NotFound desc = could not find container \"5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d\": container with ID starting with 5eeb69fc7c28355b926697dcddc815d8285c5970fc56b207dc5f577ea3e22d2d not found: ID does not exist" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.867670 4727 scope.go:117] "RemoveContainer" containerID="f4d1f8e6cc859b5c6cf1fa37d852f164ba166f4cfed768742f721e63f47cc76d" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.868702 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m" (OuterVolumeSpecName: "kube-api-access-t2c9m") pod "66fb25b1-5331-4450-87a0-9e032e6ef984" (UID: "66fb25b1-5331-4450-87a0-9e032e6ef984"). InnerVolumeSpecName "kube-api-access-t2c9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.868813 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln" (OuterVolumeSpecName: "kube-api-access-ncbln") pod "b5b4ca6c-b281-45ab-9417-d2903007a4f8" (UID: "b5b4ca6c-b281-45ab-9417-d2903007a4f8"). InnerVolumeSpecName "kube-api-access-ncbln". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.871455 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt" (OuterVolumeSpecName: "kube-api-access-m8vpt") pod "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" (UID: "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c"). InnerVolumeSpecName "kube-api-access-m8vpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.876629 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" (UID: "cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.889004 4727 scope.go:117] "RemoveContainer" containerID="303f982ce9428d3b6d6a76946d5a632b45454e1186bbb7f2f1d5795ba4f7816f" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.906050 4727 scope.go:117] "RemoveContainer" containerID="b9ee8981ff63daf4348c55af8b3a9974643c6fd61e28deb96ce8e66345f57acf" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.908982 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66fb25b1-5331-4450-87a0-9e032e6ef984" (UID: "66fb25b1-5331-4450-87a0-9e032e6ef984"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.918034 4727 scope.go:117] "RemoveContainer" containerID="de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.928867 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5b4ca6c-b281-45ab-9417-d2903007a4f8" (UID: "b5b4ca6c-b281-45ab-9417-d2903007a4f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.934819 4727 scope.go:117] "RemoveContainer" containerID="7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.960062 4727 scope.go:117] "RemoveContainer" containerID="bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963519 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2c9m\" (UniqueName: \"kubernetes.io/projected/66fb25b1-5331-4450-87a0-9e032e6ef984-kube-api-access-t2c9m\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963552 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963562 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963572 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncbln\" (UniqueName: \"kubernetes.io/projected/b5b4ca6c-b281-45ab-9417-d2903007a4f8-kube-api-access-ncbln\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963582 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963590 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66fb25b1-5331-4450-87a0-9e032e6ef984-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963599 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963607 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b4ca6c-b281-45ab-9417-d2903007a4f8-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.963615 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8vpt\" (UniqueName: \"kubernetes.io/projected/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c-kube-api-access-m8vpt\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.973786 4727 scope.go:117] "RemoveContainer" containerID="de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.974251 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d\": container with ID starting with de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d not found: ID does not exist" containerID="de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.974316 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d"} err="failed to get container status \"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d\": rpc error: code = NotFound desc = could not find container \"de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d\": container with ID starting with de103c76a435a842a1e0c245235d749b9b6c7f26cca9bf3254a50352d3bb6a2d not found: ID does not exist" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.974366 4727 scope.go:117] "RemoveContainer" containerID="7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.974813 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131\": container with ID starting with 7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131 not found: ID does not exist" containerID="7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.974855 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131"} err="failed to get container status \"7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131\": rpc error: code = NotFound desc = could not find container \"7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131\": container with ID starting with 7a27f41cd73d8e5badd42bbd6087e7295ee05573a368cbcbb0277d226304d131 not found: ID does not exist" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.974881 4727 scope.go:117] "RemoveContainer" containerID="bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2" Sep 29 10:27:31 crc kubenswrapper[4727]: E0929 10:27:31.975722 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2\": container with ID starting with bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2 not found: ID does not exist" containerID="bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2" Sep 29 10:27:31 crc kubenswrapper[4727]: I0929 10:27:31.975748 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2"} err="failed to get container status \"bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2\": rpc error: code = NotFound desc = could not find container \"bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2\": container with ID starting with bfe7819f2d71060854536454f32365acc5f6caeeb4c8fb0f124dbf81cc8710e2 not found: ID does not exist" Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.141349 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.146509 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mm76q"] Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.170498 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.173467 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tm27"] Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.176413 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:27:32 crc kubenswrapper[4727]: I0929 10:27:32.179024 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-98m5m"] Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.116074 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" path="/var/lib/kubelet/pods/66fb25b1-5331-4450-87a0-9e032e6ef984/volumes" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.116944 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" path="/var/lib/kubelet/pods/b5b4ca6c-b281-45ab-9417-d2903007a4f8/volumes" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.117768 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" path="/var/lib/kubelet/pods/cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c/volumes" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.208878 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.209195 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p52fg" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="registry-server" containerID="cri-o://bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7" gracePeriod=2 Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.542842 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.688985 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities\") pod \"eef57f61-5734-43ca-bd98-ba2e1131397e\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.690135 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gt59\" (UniqueName: \"kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59\") pod \"eef57f61-5734-43ca-bd98-ba2e1131397e\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.690203 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content\") pod \"eef57f61-5734-43ca-bd98-ba2e1131397e\" (UID: \"eef57f61-5734-43ca-bd98-ba2e1131397e\") " Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.690353 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities" (OuterVolumeSpecName: "utilities") pod "eef57f61-5734-43ca-bd98-ba2e1131397e" (UID: "eef57f61-5734-43ca-bd98-ba2e1131397e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.690856 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.695278 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59" (OuterVolumeSpecName: "kube-api-access-8gt59") pod "eef57f61-5734-43ca-bd98-ba2e1131397e" (UID: "eef57f61-5734-43ca-bd98-ba2e1131397e"). InnerVolumeSpecName "kube-api-access-8gt59". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.765171 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eef57f61-5734-43ca-bd98-ba2e1131397e" (UID: "eef57f61-5734-43ca-bd98-ba2e1131397e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.793216 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gt59\" (UniqueName: \"kubernetes.io/projected/eef57f61-5734-43ca-bd98-ba2e1131397e-kube-api-access-8gt59\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.793270 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef57f61-5734-43ca-bd98-ba2e1131397e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.858789 4727 generic.go:334] "Generic (PLEG): container finished" podID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerID="bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7" exitCode=0 Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.858857 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerDied","Data":"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7"} Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.858896 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52fg" event={"ID":"eef57f61-5734-43ca-bd98-ba2e1131397e","Type":"ContainerDied","Data":"da11f49f1a9a705d15b7f4598aaf41dc2d118c53decfa3c119a9e2bb6ace38cb"} Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.858919 4727 scope.go:117] "RemoveContainer" containerID="bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.859204 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52fg" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.874725 4727 scope.go:117] "RemoveContainer" containerID="3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.890560 4727 scope.go:117] "RemoveContainer" containerID="5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.917497 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.919453 4727 scope.go:117] "RemoveContainer" containerID="bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7" Sep 29 10:27:33 crc kubenswrapper[4727]: E0929 10:27:33.920064 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7\": container with ID starting with bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7 not found: ID does not exist" containerID="bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.920100 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7"} err="failed to get container status \"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7\": rpc error: code = NotFound desc = could not find container \"bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7\": container with ID starting with bbc28f576287ce6be695967d1302c06c5dfd732c4d486c505a1bf5847b718be7 not found: ID does not exist" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.920131 4727 scope.go:117] "RemoveContainer" containerID="3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76" Sep 29 10:27:33 crc kubenswrapper[4727]: E0929 10:27:33.921156 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76\": container with ID starting with 3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76 not found: ID does not exist" containerID="3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.921270 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76"} err="failed to get container status \"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76\": rpc error: code = NotFound desc = could not find container \"3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76\": container with ID starting with 3882eab63fef6554e7b2bd00b3d82d7479748747249ebe9b7502a21594a09d76 not found: ID does not exist" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.921363 4727 scope.go:117] "RemoveContainer" containerID="5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.921427 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p52fg"] Sep 29 10:27:33 crc kubenswrapper[4727]: E0929 10:27:33.921832 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8\": container with ID starting with 5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8 not found: ID does not exist" containerID="5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8" Sep 29 10:27:33 crc kubenswrapper[4727]: I0929 10:27:33.921890 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8"} err="failed to get container status \"5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8\": rpc error: code = NotFound desc = could not find container \"5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8\": container with ID starting with 5b4234100f5b6add9751cabbb23ad97f35906873bd43201f02cd4f68b8f623b8 not found: ID does not exist" Sep 29 10:27:35 crc kubenswrapper[4727]: I0929 10:27:35.115033 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" path="/var/lib/kubelet/pods/eef57f61-5734-43ca-bd98-ba2e1131397e/volumes" Sep 29 10:27:45 crc kubenswrapper[4727]: I0929 10:27:45.976483 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" containerID="cri-o://4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4" gracePeriod=15 Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.941772 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.955760 4727 generic.go:334] "Generic (PLEG): container finished" podID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerID="4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4" exitCode=0 Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.955817 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" event={"ID":"f1fb18d6-876c-4a19-9aab-fa0cfaccd975","Type":"ContainerDied","Data":"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4"} Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.955852 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" event={"ID":"f1fb18d6-876c-4a19-9aab-fa0cfaccd975","Type":"ContainerDied","Data":"52f41e3b6e92d4ad2bd85e390c09497738b6196431f9c43d0b3821a4d9504dac"} Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.955874 4727 scope.go:117] "RemoveContainer" containerID="4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4" Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.955904 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxzg5" Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.992754 4727 scope.go:117] "RemoveContainer" containerID="4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4" Sep 29 10:27:46 crc kubenswrapper[4727]: E0929 10:27:46.993675 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4\": container with ID starting with 4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4 not found: ID does not exist" containerID="4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4" Sep 29 10:27:46 crc kubenswrapper[4727]: I0929 10:27:46.993736 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4"} err="failed to get container status \"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4\": rpc error: code = NotFound desc = could not find container \"4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4\": container with ID starting with 4a345a26bf04da70f6f5badeda3a660c76e4266cd9afd6db28b126684ddc45d4 not found: ID does not exist" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.006804 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-78558fc4d-wgfmc"] Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007261 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007283 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007304 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007312 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007323 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007332 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007389 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007397 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007409 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ddda0d-c994-4938-89c5-96b07e7cb218" containerName="pruner" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007418 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ddda0d-c994-4938-89c5-96b07e7cb218" containerName="pruner" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007429 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007437 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007446 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007454 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007466 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007474 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007484 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007491 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007502 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007510 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007522 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007530 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="extract-utilities" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007541 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007548 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007560 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007568 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: E0929 10:27:47.007582 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007590 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="extract-content" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007744 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b4ca6c-b281-45ab-9417-d2903007a4f8" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007759 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ddda0d-c994-4938-89c5-96b07e7cb218" containerName="pruner" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007773 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="66fb25b1-5331-4450-87a0-9e032e6ef984" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007788 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1e811b-d16b-4a07-97e5-7f8d30f3eb0c" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007796 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" containerName="oauth-openshift" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.007808 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef57f61-5734-43ca-bd98-ba2e1131397e" containerName="registry-server" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.008509 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.010032 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-78558fc4d-wgfmc"] Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098607 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098666 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098700 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098720 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098809 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098836 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zqvs\" (UniqueName: \"kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098855 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098837 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.098871 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099596 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099635 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099698 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099722 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099752 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099805 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.099826 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig\") pod \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\" (UID: \"f1fb18d6-876c-4a19-9aab-fa0cfaccd975\") " Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100231 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100416 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100543 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100626 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100721 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100763 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-session\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100798 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100855 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-policies\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100893 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd4q2\" (UniqueName: \"kubernetes.io/projected/936a9d81-91c8-4893-b3b0-ffe018ed264c-kube-api-access-sd4q2\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100927 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-dir\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.100982 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-service-ca\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101030 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-error\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101052 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101225 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-router-certs\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101286 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-login\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101315 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101372 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101526 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101584 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101600 4727 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101612 4727 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.101623 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.105565 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.105726 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs" (OuterVolumeSpecName: "kube-api-access-6zqvs") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "kube-api-access-6zqvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.106330 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.106500 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.106843 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.106957 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.107325 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.107632 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.108074 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f1fb18d6-876c-4a19-9aab-fa0cfaccd975" (UID: "f1fb18d6-876c-4a19-9aab-fa0cfaccd975"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203548 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203764 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-policies\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203803 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd4q2\" (UniqueName: \"kubernetes.io/projected/936a9d81-91c8-4893-b3b0-ffe018ed264c-kube-api-access-sd4q2\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203845 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-dir\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203881 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-service-ca\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.203976 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-error\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204002 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204062 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-router-certs\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204099 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204128 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-login\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204181 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204221 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204270 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204310 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-session\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204387 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204407 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204424 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204440 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204458 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204473 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zqvs\" (UniqueName: \"kubernetes.io/projected/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-kube-api-access-6zqvs\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204488 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204501 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204514 4727 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fb18d6-876c-4a19-9aab-fa0cfaccd975-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204538 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-policies\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.204954 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.205383 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/936a9d81-91c8-4893-b3b0-ffe018ed264c-audit-dir\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.205833 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-service-ca\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.206263 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.208826 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.208852 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.209129 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.209513 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-session\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.209731 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-login\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.210329 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-router-certs\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.211893 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.212843 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/936a9d81-91c8-4893-b3b0-ffe018ed264c-v4-0-config-user-template-error\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.224299 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd4q2\" (UniqueName: \"kubernetes.io/projected/936a9d81-91c8-4893-b3b0-ffe018ed264c-kube-api-access-sd4q2\") pod \"oauth-openshift-78558fc4d-wgfmc\" (UID: \"936a9d81-91c8-4893-b3b0-ffe018ed264c\") " pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.276679 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.280704 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxzg5"] Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.335363 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.454083 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.567622 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-78558fc4d-wgfmc"] Sep 29 10:27:47 crc kubenswrapper[4727]: I0929 10:27:47.962689 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" event={"ID":"936a9d81-91c8-4893-b3b0-ffe018ed264c","Type":"ContainerStarted","Data":"0ec4fbee8b74603b711a0b5384be890475588f6ac443d4b954b5e01f13f6d49a"} Sep 29 10:27:48 crc kubenswrapper[4727]: I0929 10:27:48.970300 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" event={"ID":"936a9d81-91c8-4893-b3b0-ffe018ed264c","Type":"ContainerStarted","Data":"cecfc79f20ec09a46d728aa3c4232410b4bc1c3be3c3c2e22d15a927c10d0763"} Sep 29 10:27:48 crc kubenswrapper[4727]: I0929 10:27:48.973568 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:48 crc kubenswrapper[4727]: I0929 10:27:48.979152 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" Sep 29 10:27:49 crc kubenswrapper[4727]: I0929 10:27:49.007782 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-78558fc4d-wgfmc" podStartSLOduration=29.007756951 podStartE2EDuration="29.007756951s" podCreationTimestamp="2025-09-29 10:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:27:49.002830331 +0000 UTC m=+339.176143713" watchObservedRunningTime="2025-09-29 10:27:49.007756951 +0000 UTC m=+339.181070313" Sep 29 10:27:49 crc kubenswrapper[4727]: I0929 10:27:49.122744 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1fb18d6-876c-4a19-9aab-fa0cfaccd975" path="/var/lib/kubelet/pods/f1fb18d6-876c-4a19-9aab-fa0cfaccd975/volumes" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.330457 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.332515 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qw9xf" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="registry-server" containerID="cri-o://306ea6f8be6205841cb9a3ea5c467a82888c078ece596c702ff20cc79b9201b9" gracePeriod=30 Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.341043 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.341380 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfscf" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="registry-server" containerID="cri-o://579a705e128d65b62e9872cf42356e3bceb81b18be7e17938ba3297680c15dda" gracePeriod=30 Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.350718 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.351359 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" containerID="cri-o://2fc04b4f148353e3533d18e40f8e74abef9fd72a0a434d66b1f61813c57a45d1" gracePeriod=30 Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.368211 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.368520 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x887h" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="registry-server" containerID="cri-o://774fa49255e95b0556fd6c1381fc2fedc9a298242cbfce4d782b6041e36e3cd4" gracePeriod=30 Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.373413 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nsbj"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.374412 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.381828 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.382263 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j7lzd" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="registry-server" containerID="cri-o://7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3" gracePeriod=30 Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.384522 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nsbj"] Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.412427 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4zv9\" (UniqueName: \"kubernetes.io/projected/a9c2af09-6645-4cef-bba5-59c2b5b278d2-kube-api-access-h4zv9\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.412497 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.412539 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.513743 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4zv9\" (UniqueName: \"kubernetes.io/projected/a9c2af09-6645-4cef-bba5-59c2b5b278d2-kube-api-access-h4zv9\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.513820 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.513844 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.514927 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.527099 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9c2af09-6645-4cef-bba5-59c2b5b278d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.537059 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4zv9\" (UniqueName: \"kubernetes.io/projected/a9c2af09-6645-4cef-bba5-59c2b5b278d2-kube-api-access-h4zv9\") pod \"marketplace-operator-79b997595-4nsbj\" (UID: \"a9c2af09-6645-4cef-bba5-59c2b5b278d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.700358 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.861032 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.920595 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content\") pod \"1ed07777-5c70-4338-b9a5-dc3f12aae019\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.920784 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities\") pod \"1ed07777-5c70-4338-b9a5-dc3f12aae019\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.920845 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8xl2\" (UniqueName: \"kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2\") pod \"1ed07777-5c70-4338-b9a5-dc3f12aae019\" (UID: \"1ed07777-5c70-4338-b9a5-dc3f12aae019\") " Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.923190 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities" (OuterVolumeSpecName: "utilities") pod "1ed07777-5c70-4338-b9a5-dc3f12aae019" (UID: "1ed07777-5c70-4338-b9a5-dc3f12aae019"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:00 crc kubenswrapper[4727]: I0929 10:28:00.926707 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2" (OuterVolumeSpecName: "kube-api-access-m8xl2") pod "1ed07777-5c70-4338-b9a5-dc3f12aae019" (UID: "1ed07777-5c70-4338-b9a5-dc3f12aae019"). InnerVolumeSpecName "kube-api-access-m8xl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.024206 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.024407 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8xl2\" (UniqueName: \"kubernetes.io/projected/1ed07777-5c70-4338-b9a5-dc3f12aae019-kube-api-access-m8xl2\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.026737 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ed07777-5c70-4338-b9a5-dc3f12aae019" (UID: "1ed07777-5c70-4338-b9a5-dc3f12aae019"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.057997 4727 generic.go:334] "Generic (PLEG): container finished" podID="24076974-7055-499b-92f6-16ed1fd24121" containerID="306ea6f8be6205841cb9a3ea5c467a82888c078ece596c702ff20cc79b9201b9" exitCode=0 Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.058144 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerDied","Data":"306ea6f8be6205841cb9a3ea5c467a82888c078ece596c702ff20cc79b9201b9"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.062379 4727 generic.go:334] "Generic (PLEG): container finished" podID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerID="774fa49255e95b0556fd6c1381fc2fedc9a298242cbfce4d782b6041e36e3cd4" exitCode=0 Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.062474 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerDied","Data":"774fa49255e95b0556fd6c1381fc2fedc9a298242cbfce4d782b6041e36e3cd4"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.065426 4727 generic.go:334] "Generic (PLEG): container finished" podID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerID="7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3" exitCode=0 Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.065490 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerDied","Data":"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.065569 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7lzd" event={"ID":"1ed07777-5c70-4338-b9a5-dc3f12aae019","Type":"ContainerDied","Data":"b4f81a5137efc6b079e5526ed97415432b72bca1e06954b08f1c7e4d4e68c73c"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.065595 4727 scope.go:117] "RemoveContainer" containerID="7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.065518 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7lzd" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.071560 4727 generic.go:334] "Generic (PLEG): container finished" podID="73a9e11c-a13e-4167-9744-d7427e583d89" containerID="2fc04b4f148353e3533d18e40f8e74abef9fd72a0a434d66b1f61813c57a45d1" exitCode=0 Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.071620 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" event={"ID":"73a9e11c-a13e-4167-9744-d7427e583d89","Type":"ContainerDied","Data":"2fc04b4f148353e3533d18e40f8e74abef9fd72a0a434d66b1f61813c57a45d1"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.073670 4727 generic.go:334] "Generic (PLEG): container finished" podID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerID="579a705e128d65b62e9872cf42356e3bceb81b18be7e17938ba3297680c15dda" exitCode=0 Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.073704 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerDied","Data":"579a705e128d65b62e9872cf42356e3bceb81b18be7e17938ba3297680c15dda"} Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.106755 4727 scope.go:117] "RemoveContainer" containerID="4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.124517 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.124567 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j7lzd"] Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.125505 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed07777-5c70-4338-b9a5-dc3f12aae019-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.135523 4727 scope.go:117] "RemoveContainer" containerID="0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.182977 4727 scope.go:117] "RemoveContainer" containerID="7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3" Sep 29 10:28:01 crc kubenswrapper[4727]: E0929 10:28:01.185646 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3\": container with ID starting with 7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3 not found: ID does not exist" containerID="7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.185773 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3"} err="failed to get container status \"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3\": rpc error: code = NotFound desc = could not find container \"7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3\": container with ID starting with 7fb2a7ae3e41edf1fab72b2599148ddfddfa0d8b440c33da5299c13577758ef3 not found: ID does not exist" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.185858 4727 scope.go:117] "RemoveContainer" containerID="4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38" Sep 29 10:28:01 crc kubenswrapper[4727]: E0929 10:28:01.186270 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38\": container with ID starting with 4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38 not found: ID does not exist" containerID="4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.186295 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38"} err="failed to get container status \"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38\": rpc error: code = NotFound desc = could not find container \"4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38\": container with ID starting with 4d384337817dab24c6bfb198af7b01a9a156db47dcf313ffd1e6016ee7dd4a38 not found: ID does not exist" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.186315 4727 scope.go:117] "RemoveContainer" containerID="0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e" Sep 29 10:28:01 crc kubenswrapper[4727]: E0929 10:28:01.186846 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e\": container with ID starting with 0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e not found: ID does not exist" containerID="0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.186882 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e"} err="failed to get container status \"0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e\": rpc error: code = NotFound desc = could not find container \"0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e\": container with ID starting with 0c0074f7df2dde39d377c1d91573500c4259b379b1eed19c77c51315168c295e not found: ID does not exist" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.232542 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nsbj"] Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.264981 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.328962 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content\") pod \"24076974-7055-499b-92f6-16ed1fd24121\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.329126 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities\") pod \"24076974-7055-499b-92f6-16ed1fd24121\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.329164 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7llc\" (UniqueName: \"kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc\") pod \"24076974-7055-499b-92f6-16ed1fd24121\" (UID: \"24076974-7055-499b-92f6-16ed1fd24121\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.331124 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities" (OuterVolumeSpecName: "utilities") pod "24076974-7055-499b-92f6-16ed1fd24121" (UID: "24076974-7055-499b-92f6-16ed1fd24121"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.337265 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc" (OuterVolumeSpecName: "kube-api-access-x7llc") pod "24076974-7055-499b-92f6-16ed1fd24121" (UID: "24076974-7055-499b-92f6-16ed1fd24121"). InnerVolumeSpecName "kube-api-access-x7llc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.339681 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.364962 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.407993 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.418232 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24076974-7055-499b-92f6-16ed1fd24121" (UID: "24076974-7055-499b-92f6-16ed1fd24121"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.441351 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.441413 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7llc\" (UniqueName: \"kubernetes.io/projected/24076974-7055-499b-92f6-16ed1fd24121-kube-api-access-x7llc\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.442143 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24076974-7055-499b-92f6-16ed1fd24121-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.542786 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph95m\" (UniqueName: \"kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m\") pod \"73a9e11c-a13e-4167-9744-d7427e583d89\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.542854 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities\") pod \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.542907 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content\") pod \"03310bb6-f32c-48a0-a2df-a34b133b12c3\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.542952 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics\") pod \"73a9e11c-a13e-4167-9744-d7427e583d89\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.542975 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca\") pod \"73a9e11c-a13e-4167-9744-d7427e583d89\" (UID: \"73a9e11c-a13e-4167-9744-d7427e583d89\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.543037 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ql58\" (UniqueName: \"kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58\") pod \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.543079 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content\") pod \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\" (UID: \"d2dcef0d-c089-4c5a-967b-19beeaceb1de\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.543111 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities\") pod \"03310bb6-f32c-48a0-a2df-a34b133b12c3\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.543143 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4vh4\" (UniqueName: \"kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4\") pod \"03310bb6-f32c-48a0-a2df-a34b133b12c3\" (UID: \"03310bb6-f32c-48a0-a2df-a34b133b12c3\") " Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.545069 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities" (OuterVolumeSpecName: "utilities") pod "d2dcef0d-c089-4c5a-967b-19beeaceb1de" (UID: "d2dcef0d-c089-4c5a-967b-19beeaceb1de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.545386 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities" (OuterVolumeSpecName: "utilities") pod "03310bb6-f32c-48a0-a2df-a34b133b12c3" (UID: "03310bb6-f32c-48a0-a2df-a34b133b12c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.545499 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "73a9e11c-a13e-4167-9744-d7427e583d89" (UID: "73a9e11c-a13e-4167-9744-d7427e583d89"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.548130 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "73a9e11c-a13e-4167-9744-d7427e583d89" (UID: "73a9e11c-a13e-4167-9744-d7427e583d89"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.559785 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m" (OuterVolumeSpecName: "kube-api-access-ph95m") pod "73a9e11c-a13e-4167-9744-d7427e583d89" (UID: "73a9e11c-a13e-4167-9744-d7427e583d89"). InnerVolumeSpecName "kube-api-access-ph95m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.563257 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4" (OuterVolumeSpecName: "kube-api-access-r4vh4") pod "03310bb6-f32c-48a0-a2df-a34b133b12c3" (UID: "03310bb6-f32c-48a0-a2df-a34b133b12c3"). InnerVolumeSpecName "kube-api-access-r4vh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.566438 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03310bb6-f32c-48a0-a2df-a34b133b12c3" (UID: "03310bb6-f32c-48a0-a2df-a34b133b12c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.566921 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58" (OuterVolumeSpecName: "kube-api-access-9ql58") pod "d2dcef0d-c089-4c5a-967b-19beeaceb1de" (UID: "d2dcef0d-c089-4c5a-967b-19beeaceb1de"). InnerVolumeSpecName "kube-api-access-9ql58". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.592261 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2dcef0d-c089-4c5a-967b-19beeaceb1de" (UID: "d2dcef0d-c089-4c5a-967b-19beeaceb1de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.644995 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645072 4727 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645090 4727 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a9e11c-a13e-4167-9744-d7427e583d89-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645103 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ql58\" (UniqueName: \"kubernetes.io/projected/d2dcef0d-c089-4c5a-967b-19beeaceb1de-kube-api-access-9ql58\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645117 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645130 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03310bb6-f32c-48a0-a2df-a34b133b12c3-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645140 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4vh4\" (UniqueName: \"kubernetes.io/projected/03310bb6-f32c-48a0-a2df-a34b133b12c3-kube-api-access-r4vh4\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645150 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph95m\" (UniqueName: \"kubernetes.io/projected/73a9e11c-a13e-4167-9744-d7427e583d89-kube-api-access-ph95m\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:01 crc kubenswrapper[4727]: I0929 10:28:01.645167 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2dcef0d-c089-4c5a-967b-19beeaceb1de-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.081968 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.082465 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-65ccr" event={"ID":"73a9e11c-a13e-4167-9744-d7427e583d89","Type":"ContainerDied","Data":"a4fd05c74d62fc64a31b2c813e9aae1011bbd26c7633a5a86e201b37e726bfc9"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.082536 4727 scope.go:117] "RemoveContainer" containerID="2fc04b4f148353e3533d18e40f8e74abef9fd72a0a434d66b1f61813c57a45d1" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.084950 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfscf" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.084902 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfscf" event={"ID":"d2dcef0d-c089-4c5a-967b-19beeaceb1de","Type":"ContainerDied","Data":"ec3a2562d11ba5955b62aaad0a400e79b68f2811eab9503a9db4ef5d6e65c520"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.087539 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x887h" event={"ID":"03310bb6-f32c-48a0-a2df-a34b133b12c3","Type":"ContainerDied","Data":"6bd12842e245b4c27d8910538a75cc9f59ec45013ec5daf91f742c3c13b32b9d"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.087570 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x887h" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.088877 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" event={"ID":"a9c2af09-6645-4cef-bba5-59c2b5b278d2","Type":"ContainerStarted","Data":"987f2fe9f0f70783b4d1a6f4f0ac0a4f27796dd206f9a5104da263164eae040a"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.088911 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" event={"ID":"a9c2af09-6645-4cef-bba5-59c2b5b278d2","Type":"ContainerStarted","Data":"3acaf3e8c031b69b671dfba1237ff771fa174c5037e2de58f9c281f5424b1e57"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.090200 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.096456 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9xf" event={"ID":"24076974-7055-499b-92f6-16ed1fd24121","Type":"ContainerDied","Data":"a4bc973e3d99f21e67ef2d7af40f5187a47ec9ee492677cdf6904de986f1c183"} Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.096545 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9xf" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.097483 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.107151 4727 scope.go:117] "RemoveContainer" containerID="579a705e128d65b62e9872cf42356e3bceb81b18be7e17938ba3297680c15dda" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.135522 4727 scope.go:117] "RemoveContainer" containerID="0d306692a506ead0120b9a11a69496374b269a0d692c90515ccfc05ef6c59f43" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.145543 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4nsbj" podStartSLOduration=2.14552177 podStartE2EDuration="2.14552177s" podCreationTimestamp="2025-09-29 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:28:02.128126043 +0000 UTC m=+352.301439415" watchObservedRunningTime="2025-09-29 10:28:02.14552177 +0000 UTC m=+352.318835132" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.155065 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.163686 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfscf"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.164317 4727 scope.go:117] "RemoveContainer" containerID="2de2baca6fb82adccfcd528d514b646f7697bc0ad4a9d0663ca1d2588e09d559" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.177419 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.182442 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x887h"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.187204 4727 scope.go:117] "RemoveContainer" containerID="774fa49255e95b0556fd6c1381fc2fedc9a298242cbfce4d782b6041e36e3cd4" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.211682 4727 scope.go:117] "RemoveContainer" containerID="1a59db951f74e7ba6e61b3146ef587284a21b970641c30fc4463aa5c554463cf" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.258087 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.271129 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qw9xf"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.281615 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.286062 4727 scope.go:117] "RemoveContainer" containerID="e32941b33f5599311c0b02c3c4b9646a9e3f4dd4bf87ebcea070ee203b3082c5" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.286090 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-65ccr"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.301330 4727 scope.go:117] "RemoveContainer" containerID="306ea6f8be6205841cb9a3ea5c467a82888c078ece596c702ff20cc79b9201b9" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.316019 4727 scope.go:117] "RemoveContainer" containerID="d830a17a298fc873f4cae63e5ba32484de9d8855768f9eb12a82bc5ab82d9eea" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.334644 4727 scope.go:117] "RemoveContainer" containerID="9332668c92e54860a4354fcf3cac3de19118c73dfd051e966c6d8767f5930470" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.751891 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9bhmq"] Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752151 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752167 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752177 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752183 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752194 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752200 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752208 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752214 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752223 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752229 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752238 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752246 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752255 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752265 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752280 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752291 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752302 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752309 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752322 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752328 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752356 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752365 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="extract-content" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752375 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752383 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" Sep 29 10:28:02 crc kubenswrapper[4727]: E0929 10:28:02.752394 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752402 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="extract-utilities" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752526 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" containerName="marketplace-operator" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752541 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="24076974-7055-499b-92f6-16ed1fd24121" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752550 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752561 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.752573 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" containerName="registry-server" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.753456 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.756143 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.760834 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bhmq"] Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.764950 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg77f\" (UniqueName: \"kubernetes.io/projected/3410083e-6ab2-4674-8544-9ed975c2db00-kube-api-access-lg77f\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.765029 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-utilities\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.765066 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-catalog-content\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.866468 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg77f\" (UniqueName: \"kubernetes.io/projected/3410083e-6ab2-4674-8544-9ed975c2db00-kube-api-access-lg77f\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.866552 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-utilities\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.866589 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-catalog-content\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.867287 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-catalog-content\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.867432 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3410083e-6ab2-4674-8544-9ed975c2db00-utilities\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:02 crc kubenswrapper[4727]: I0929 10:28:02.889258 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg77f\" (UniqueName: \"kubernetes.io/projected/3410083e-6ab2-4674-8544-9ed975c2db00-kube-api-access-lg77f\") pod \"certified-operators-9bhmq\" (UID: \"3410083e-6ab2-4674-8544-9ed975c2db00\") " pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.073523 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.119529 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03310bb6-f32c-48a0-a2df-a34b133b12c3" path="/var/lib/kubelet/pods/03310bb6-f32c-48a0-a2df-a34b133b12c3/volumes" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.120423 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed07777-5c70-4338-b9a5-dc3f12aae019" path="/var/lib/kubelet/pods/1ed07777-5c70-4338-b9a5-dc3f12aae019/volumes" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.121043 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24076974-7055-499b-92f6-16ed1fd24121" path="/var/lib/kubelet/pods/24076974-7055-499b-92f6-16ed1fd24121/volumes" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.122232 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73a9e11c-a13e-4167-9744-d7427e583d89" path="/var/lib/kubelet/pods/73a9e11c-a13e-4167-9744-d7427e583d89/volumes" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.122751 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2dcef0d-c089-4c5a-967b-19beeaceb1de" path="/var/lib/kubelet/pods/d2dcef0d-c089-4c5a-967b-19beeaceb1de/volumes" Sep 29 10:28:03 crc kubenswrapper[4727]: I0929 10:28:03.353644 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bhmq"] Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.128763 4727 generic.go:334] "Generic (PLEG): container finished" podID="3410083e-6ab2-4674-8544-9ed975c2db00" containerID="1196428dd8ac47b2127e3bb190f74941dced18f46fa1f09440b8ba1856d24f99" exitCode=0 Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.128898 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bhmq" event={"ID":"3410083e-6ab2-4674-8544-9ed975c2db00","Type":"ContainerDied","Data":"1196428dd8ac47b2127e3bb190f74941dced18f46fa1f09440b8ba1856d24f99"} Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.130845 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bhmq" event={"ID":"3410083e-6ab2-4674-8544-9ed975c2db00","Type":"ContainerStarted","Data":"eb664cbf25444d30e7a59f12c3a3a0dad95372caa09a0b3304c5efeb08f2eae0"} Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.554464 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lfbq8"] Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.555880 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.559020 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.568774 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfbq8"] Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.695381 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-utilities\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.696063 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk4xb\" (UniqueName: \"kubernetes.io/projected/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-kube-api-access-fk4xb\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.696164 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-catalog-content\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.797043 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-catalog-content\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.797157 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-utilities\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.797236 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk4xb\" (UniqueName: \"kubernetes.io/projected/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-kube-api-access-fk4xb\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.797756 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-catalog-content\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.797815 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-utilities\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.827732 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk4xb\" (UniqueName: \"kubernetes.io/projected/eddf2895-0cde-4925-8ddb-98bdd9fd4f64-kube-api-access-fk4xb\") pod \"community-operators-lfbq8\" (UID: \"eddf2895-0cde-4925-8ddb-98bdd9fd4f64\") " pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:04 crc kubenswrapper[4727]: I0929 10:28:04.886807 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.097250 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfbq8"] Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.143767 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfbq8" event={"ID":"eddf2895-0cde-4925-8ddb-98bdd9fd4f64","Type":"ContainerStarted","Data":"43f406e12a612bae2dac78004a1f67aa3a906fc2d266322d8c39d87db8c9922d"} Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.144252 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.145868 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.148861 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.149362 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bhmq" event={"ID":"3410083e-6ab2-4674-8544-9ed975c2db00","Type":"ContainerStarted","Data":"9e85a32da75689264aaa77551a86393e85d565adc9ce51ecc48df07bbdfc65ba"} Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.156205 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.304433 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.304597 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.304625 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44hzs\" (UniqueName: \"kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.406009 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.406223 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.406252 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44hzs\" (UniqueName: \"kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.406603 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.406803 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.428496 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44hzs\" (UniqueName: \"kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs\") pod \"redhat-operators-9zv28\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.485109 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:05 crc kubenswrapper[4727]: I0929 10:28:05.695843 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:28:05 crc kubenswrapper[4727]: W0929 10:28:05.706812 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeddd788b_0bbb_4ffc_a3de_0dbfbb39e21c.slice/crio-c11e45aced53f8243a5b7cab9b56974627cd78c69120cb08f69068b7284db750 WatchSource:0}: Error finding container c11e45aced53f8243a5b7cab9b56974627cd78c69120cb08f69068b7284db750: Status 404 returned error can't find the container with id c11e45aced53f8243a5b7cab9b56974627cd78c69120cb08f69068b7284db750 Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.157137 4727 generic.go:334] "Generic (PLEG): container finished" podID="3410083e-6ab2-4674-8544-9ed975c2db00" containerID="9e85a32da75689264aaa77551a86393e85d565adc9ce51ecc48df07bbdfc65ba" exitCode=0 Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.157249 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bhmq" event={"ID":"3410083e-6ab2-4674-8544-9ed975c2db00","Type":"ContainerDied","Data":"9e85a32da75689264aaa77551a86393e85d565adc9ce51ecc48df07bbdfc65ba"} Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.163840 4727 generic.go:334] "Generic (PLEG): container finished" podID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerID="f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d" exitCode=0 Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.163960 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerDied","Data":"f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d"} Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.164001 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerStarted","Data":"c11e45aced53f8243a5b7cab9b56974627cd78c69120cb08f69068b7284db750"} Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.169449 4727 generic.go:334] "Generic (PLEG): container finished" podID="eddf2895-0cde-4925-8ddb-98bdd9fd4f64" containerID="c1f5a37854366390045907058b1ea293398af30b0c4aa73fdca0c0c0d09f6083" exitCode=0 Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.170032 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfbq8" event={"ID":"eddf2895-0cde-4925-8ddb-98bdd9fd4f64","Type":"ContainerDied","Data":"c1f5a37854366390045907058b1ea293398af30b0c4aa73fdca0c0c0d09f6083"} Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.949210 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hp9xv"] Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.950590 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.953236 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 10:28:06 crc kubenswrapper[4727]: I0929 10:28:06.974058 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp9xv"] Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.128328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nf99\" (UniqueName: \"kubernetes.io/projected/6e4c7812-66fe-4e23-bb8e-91b2719840f5-kube-api-access-5nf99\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.128853 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-utilities\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.128984 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-catalog-content\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.178788 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfbq8" event={"ID":"eddf2895-0cde-4925-8ddb-98bdd9fd4f64","Type":"ContainerStarted","Data":"ea23a7f4e27d4258cee15e205bf090dd52a7d4c1d96d8dfe62494bd511997ec9"} Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.230481 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-utilities\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.230789 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-catalog-content\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.230974 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-utilities\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.231108 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nf99\" (UniqueName: \"kubernetes.io/projected/6e4c7812-66fe-4e23-bb8e-91b2719840f5-kube-api-access-5nf99\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.231474 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4c7812-66fe-4e23-bb8e-91b2719840f5-catalog-content\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.254254 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nf99\" (UniqueName: \"kubernetes.io/projected/6e4c7812-66fe-4e23-bb8e-91b2719840f5-kube-api-access-5nf99\") pod \"redhat-marketplace-hp9xv\" (UID: \"6e4c7812-66fe-4e23-bb8e-91b2719840f5\") " pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.278047 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:07 crc kubenswrapper[4727]: I0929 10:28:07.522287 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp9xv"] Sep 29 10:28:07 crc kubenswrapper[4727]: W0929 10:28:07.529275 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e4c7812_66fe_4e23_bb8e_91b2719840f5.slice/crio-6cf4d7249252ae3a426315d89e8278a5e63b190315c633b66f2499c33a5a159a WatchSource:0}: Error finding container 6cf4d7249252ae3a426315d89e8278a5e63b190315c633b66f2499c33a5a159a: Status 404 returned error can't find the container with id 6cf4d7249252ae3a426315d89e8278a5e63b190315c633b66f2499c33a5a159a Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.192813 4727 generic.go:334] "Generic (PLEG): container finished" podID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerID="03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3" exitCode=0 Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.193266 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerDied","Data":"03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3"} Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.200555 4727 generic.go:334] "Generic (PLEG): container finished" podID="eddf2895-0cde-4925-8ddb-98bdd9fd4f64" containerID="ea23a7f4e27d4258cee15e205bf090dd52a7d4c1d96d8dfe62494bd511997ec9" exitCode=0 Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.201317 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfbq8" event={"ID":"eddf2895-0cde-4925-8ddb-98bdd9fd4f64","Type":"ContainerDied","Data":"ea23a7f4e27d4258cee15e205bf090dd52a7d4c1d96d8dfe62494bd511997ec9"} Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.211899 4727 generic.go:334] "Generic (PLEG): container finished" podID="6e4c7812-66fe-4e23-bb8e-91b2719840f5" containerID="73185c2e510b240a695fca60aec3b633943bb34b84246a2d5c80a5d08498993a" exitCode=0 Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.212012 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp9xv" event={"ID":"6e4c7812-66fe-4e23-bb8e-91b2719840f5","Type":"ContainerDied","Data":"73185c2e510b240a695fca60aec3b633943bb34b84246a2d5c80a5d08498993a"} Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.212049 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp9xv" event={"ID":"6e4c7812-66fe-4e23-bb8e-91b2719840f5","Type":"ContainerStarted","Data":"6cf4d7249252ae3a426315d89e8278a5e63b190315c633b66f2499c33a5a159a"} Sep 29 10:28:08 crc kubenswrapper[4727]: I0929 10:28:08.223884 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bhmq" event={"ID":"3410083e-6ab2-4674-8544-9ed975c2db00","Type":"ContainerStarted","Data":"c94a88eaa3b029374f896e4af722c7dc82e900395e44000ee93ebadaf6a3543d"} Sep 29 10:28:09 crc kubenswrapper[4727]: I0929 10:28:09.235175 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerStarted","Data":"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280"} Sep 29 10:28:09 crc kubenswrapper[4727]: I0929 10:28:09.238212 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfbq8" event={"ID":"eddf2895-0cde-4925-8ddb-98bdd9fd4f64","Type":"ContainerStarted","Data":"98ffd870f891aed5e8fc857c9e7b5ec7253e9f3dd5f5c6eebbbcad69d283797e"} Sep 29 10:28:09 crc kubenswrapper[4727]: I0929 10:28:09.257848 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9bhmq" podStartSLOduration=4.269036503 podStartE2EDuration="7.257832169s" podCreationTimestamp="2025-09-29 10:28:02 +0000 UTC" firstStartedPulling="2025-09-29 10:28:04.130908163 +0000 UTC m=+354.304221525" lastFinishedPulling="2025-09-29 10:28:07.119703829 +0000 UTC m=+357.293017191" observedRunningTime="2025-09-29 10:28:08.28199659 +0000 UTC m=+358.455309952" watchObservedRunningTime="2025-09-29 10:28:09.257832169 +0000 UTC m=+359.431145531" Sep 29 10:28:09 crc kubenswrapper[4727]: I0929 10:28:09.259656 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9zv28" podStartSLOduration=1.438787265 podStartE2EDuration="4.259648546s" podCreationTimestamp="2025-09-29 10:28:05 +0000 UTC" firstStartedPulling="2025-09-29 10:28:06.165966692 +0000 UTC m=+356.339280054" lastFinishedPulling="2025-09-29 10:28:08.986827973 +0000 UTC m=+359.160141335" observedRunningTime="2025-09-29 10:28:09.254945643 +0000 UTC m=+359.428259005" watchObservedRunningTime="2025-09-29 10:28:09.259648546 +0000 UTC m=+359.432961908" Sep 29 10:28:11 crc kubenswrapper[4727]: I0929 10:28:11.253082 4727 generic.go:334] "Generic (PLEG): container finished" podID="6e4c7812-66fe-4e23-bb8e-91b2719840f5" containerID="6404c9c348176bfd770de8339dc8ef981abd5a812907410a81e8120eee6bcd2a" exitCode=0 Sep 29 10:28:11 crc kubenswrapper[4727]: I0929 10:28:11.253699 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp9xv" event={"ID":"6e4c7812-66fe-4e23-bb8e-91b2719840f5","Type":"ContainerDied","Data":"6404c9c348176bfd770de8339dc8ef981abd5a812907410a81e8120eee6bcd2a"} Sep 29 10:28:11 crc kubenswrapper[4727]: I0929 10:28:11.280049 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lfbq8" podStartSLOduration=4.700929706 podStartE2EDuration="7.280030269s" podCreationTimestamp="2025-09-29 10:28:04 +0000 UTC" firstStartedPulling="2025-09-29 10:28:06.172071873 +0000 UTC m=+356.345385235" lastFinishedPulling="2025-09-29 10:28:08.751172436 +0000 UTC m=+358.924485798" observedRunningTime="2025-09-29 10:28:09.279317773 +0000 UTC m=+359.452631145" watchObservedRunningTime="2025-09-29 10:28:11.280030269 +0000 UTC m=+361.453343631" Sep 29 10:28:12 crc kubenswrapper[4727]: I0929 10:28:12.262730 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp9xv" event={"ID":"6e4c7812-66fe-4e23-bb8e-91b2719840f5","Type":"ContainerStarted","Data":"8ca576beb1c95cca78aff858c64800b4f1a4101b447e902a1fd15525ecbca6d0"} Sep 29 10:28:12 crc kubenswrapper[4727]: I0929 10:28:12.287616 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hp9xv" podStartSLOduration=2.780830437 podStartE2EDuration="6.287588452s" podCreationTimestamp="2025-09-29 10:28:06 +0000 UTC" firstStartedPulling="2025-09-29 10:28:08.222538717 +0000 UTC m=+358.395852079" lastFinishedPulling="2025-09-29 10:28:11.729296732 +0000 UTC m=+361.902610094" observedRunningTime="2025-09-29 10:28:12.286975966 +0000 UTC m=+362.460289328" watchObservedRunningTime="2025-09-29 10:28:12.287588452 +0000 UTC m=+362.460901814" Sep 29 10:28:13 crc kubenswrapper[4727]: I0929 10:28:13.074407 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:13 crc kubenswrapper[4727]: I0929 10:28:13.074795 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:13 crc kubenswrapper[4727]: I0929 10:28:13.130136 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:13 crc kubenswrapper[4727]: I0929 10:28:13.314884 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9bhmq" Sep 29 10:28:14 crc kubenswrapper[4727]: I0929 10:28:14.887119 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:14 crc kubenswrapper[4727]: I0929 10:28:14.887650 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:14 crc kubenswrapper[4727]: I0929 10:28:14.936499 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:15 crc kubenswrapper[4727]: I0929 10:28:15.324308 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lfbq8" Sep 29 10:28:15 crc kubenswrapper[4727]: I0929 10:28:15.485859 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:15 crc kubenswrapper[4727]: I0929 10:28:15.486368 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:15 crc kubenswrapper[4727]: I0929 10:28:15.524721 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:16 crc kubenswrapper[4727]: I0929 10:28:16.334622 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:28:17 crc kubenswrapper[4727]: I0929 10:28:17.278437 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:17 crc kubenswrapper[4727]: I0929 10:28:17.278507 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:17 crc kubenswrapper[4727]: I0929 10:28:17.321325 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:17 crc kubenswrapper[4727]: I0929 10:28:17.359721 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hp9xv" Sep 29 10:28:19 crc kubenswrapper[4727]: I0929 10:28:19.246778 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:28:19 crc kubenswrapper[4727]: I0929 10:28:19.246855 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:28:49 crc kubenswrapper[4727]: I0929 10:28:49.246601 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:28:49 crc kubenswrapper[4727]: I0929 10:28:49.247536 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.431619 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9tfv5"] Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.433441 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.447422 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9tfv5"] Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521420 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-tls\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521478 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-certificates\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521507 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521529 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvwth\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-kube-api-access-cvwth\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521548 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9adac727-d3a3-4038-af8f-54b6a404bc65-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521567 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-bound-sa-token\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521586 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-trusted-ca\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.521619 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9adac727-d3a3-4038-af8f-54b6a404bc65-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.546518 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.622855 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-bound-sa-token\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.622933 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-trusted-ca\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.622983 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9adac727-d3a3-4038-af8f-54b6a404bc65-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.623015 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-tls\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.623055 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-certificates\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.623081 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvwth\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-kube-api-access-cvwth\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.623101 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9adac727-d3a3-4038-af8f-54b6a404bc65-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.624138 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9adac727-d3a3-4038-af8f-54b6a404bc65-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.624999 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-trusted-ca\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.625028 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-certificates\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.633538 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9adac727-d3a3-4038-af8f-54b6a404bc65-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.633590 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-registry-tls\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.639973 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvwth\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-kube-api-access-cvwth\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.641646 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9adac727-d3a3-4038-af8f-54b6a404bc65-bound-sa-token\") pod \"image-registry-66df7c8f76-9tfv5\" (UID: \"9adac727-d3a3-4038-af8f-54b6a404bc65\") " pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.757150 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:05 crc kubenswrapper[4727]: I0929 10:29:05.991653 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9tfv5"] Sep 29 10:29:06 crc kubenswrapper[4727]: I0929 10:29:06.598680 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" event={"ID":"9adac727-d3a3-4038-af8f-54b6a404bc65","Type":"ContainerStarted","Data":"6d9a2e6c1310f92e40c2bcf75046f9e329bd0dda4029e9bda05418db1bc84cef"} Sep 29 10:29:06 crc kubenswrapper[4727]: I0929 10:29:06.598744 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" event={"ID":"9adac727-d3a3-4038-af8f-54b6a404bc65","Type":"ContainerStarted","Data":"386cae7bc130b3c15a65a58b0ef68f9b86bb30dcbcd2739bcd14e2cd4ee7dced"} Sep 29 10:29:06 crc kubenswrapper[4727]: I0929 10:29:06.598874 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:06 crc kubenswrapper[4727]: I0929 10:29:06.620554 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" podStartSLOduration=1.6205311789999999 podStartE2EDuration="1.620531179s" podCreationTimestamp="2025-09-29 10:29:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:29:06.618832495 +0000 UTC m=+416.792145857" watchObservedRunningTime="2025-09-29 10:29:06.620531179 +0000 UTC m=+416.793844541" Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.247219 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.247667 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.247713 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.248207 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.248265 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc" gracePeriod=600 Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.674554 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc" exitCode=0 Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.675168 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc"} Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.675288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16"} Sep 29 10:29:19 crc kubenswrapper[4727]: I0929 10:29:19.675316 4727 scope.go:117] "RemoveContainer" containerID="896a54f6f25637f65085212e3f1cf22bd4e5250996cef646a2e4697208eefc47" Sep 29 10:29:25 crc kubenswrapper[4727]: I0929 10:29:25.765016 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9tfv5" Sep 29 10:29:25 crc kubenswrapper[4727]: I0929 10:29:25.833901 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:29:50 crc kubenswrapper[4727]: I0929 10:29:50.874096 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerName="registry" containerID="cri-o://7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9" gracePeriod=30 Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.599456 4727 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-nvqqb container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.35:5000/healthz\": dial tcp 10.217.0.35:5000: connect: connection refused" start-of-body= Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.600830 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.35:5000/healthz\": dial tcp 10.217.0.35:5000: connect: connection refused" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.744974 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806383 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806446 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrk6v\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806470 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806492 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806524 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806585 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806615 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.806683 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted\") pod \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\" (UID: \"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0\") " Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.813501 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.814154 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.828479 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.831689 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.833445 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.833474 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v" (OuterVolumeSpecName: "kube-api-access-jrk6v") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "kube-api-access-jrk6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.834204 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.849158 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" (UID: "5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.875870 4727 generic.go:334] "Generic (PLEG): container finished" podID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerID="7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9" exitCode=0 Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.875922 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" event={"ID":"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0","Type":"ContainerDied","Data":"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9"} Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.875961 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" event={"ID":"5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0","Type":"ContainerDied","Data":"990ddb3b196207b633cbb2901f3d7842d3d2a45e20538da56fcb3c5c40333519"} Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.875981 4727 scope.go:117] "RemoveContainer" containerID="7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.875971 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nvqqb" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.894847 4727 scope.go:117] "RemoveContainer" containerID="7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9" Sep 29 10:29:51 crc kubenswrapper[4727]: E0929 10:29:51.895572 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9\": container with ID starting with 7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9 not found: ID does not exist" containerID="7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.895660 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9"} err="failed to get container status \"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9\": rpc error: code = NotFound desc = could not find container \"7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9\": container with ID starting with 7d61bc261ea3da3135da4a843a2b28c6ac929e0a2fa4e3778e61d34858f391a9 not found: ID does not exist" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.904578 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.907708 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nvqqb"] Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908084 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrk6v\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-kube-api-access-jrk6v\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908226 4727 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908299 4727 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908409 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908477 4727 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908543 4727 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:51 crc kubenswrapper[4727]: I0929 10:29:51.908623 4727 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 10:29:53 crc kubenswrapper[4727]: I0929 10:29:53.123183 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" path="/var/lib/kubelet/pods/5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0/volumes" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.128453 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx"] Sep 29 10:30:00 crc kubenswrapper[4727]: E0929 10:30:00.129233 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerName="registry" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.129248 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerName="registry" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.129382 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0aeaf8-e15c-4bb5-911f-f55b754d1ae0" containerName="registry" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.129859 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.133362 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.133379 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.140102 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx"] Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.219932 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.220005 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9lt4\" (UniqueName: \"kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.221044 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.322347 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.322423 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.322463 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9lt4\" (UniqueName: \"kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.323514 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.330372 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.342374 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9lt4\" (UniqueName: \"kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4\") pod \"collect-profiles-29319030-52fpx\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.451878 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.653186 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx"] Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.930083 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" event={"ID":"fac7e10c-b919-4e7b-b2b2-4801fe56643b","Type":"ContainerStarted","Data":"bd9ec1520f3854213e8eaa06cfedbadbc7843e71065724f2db78cb0ee994a350"} Sep 29 10:30:00 crc kubenswrapper[4727]: I0929 10:30:00.930140 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" event={"ID":"fac7e10c-b919-4e7b-b2b2-4801fe56643b","Type":"ContainerStarted","Data":"33b6aa94609adcb06d33108015a2b18f2af0c66215e9977186cf847890582a1f"} Sep 29 10:30:01 crc kubenswrapper[4727]: I0929 10:30:01.938450 4727 generic.go:334] "Generic (PLEG): container finished" podID="fac7e10c-b919-4e7b-b2b2-4801fe56643b" containerID="bd9ec1520f3854213e8eaa06cfedbadbc7843e71065724f2db78cb0ee994a350" exitCode=0 Sep 29 10:30:01 crc kubenswrapper[4727]: I0929 10:30:01.938509 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" event={"ID":"fac7e10c-b919-4e7b-b2b2-4801fe56643b","Type":"ContainerDied","Data":"bd9ec1520f3854213e8eaa06cfedbadbc7843e71065724f2db78cb0ee994a350"} Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.158731 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.245715 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9lt4\" (UniqueName: \"kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4\") pod \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.245800 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume\") pod \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.245833 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume\") pod \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\" (UID: \"fac7e10c-b919-4e7b-b2b2-4801fe56643b\") " Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.247514 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume" (OuterVolumeSpecName: "config-volume") pod "fac7e10c-b919-4e7b-b2b2-4801fe56643b" (UID: "fac7e10c-b919-4e7b-b2b2-4801fe56643b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.251008 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fac7e10c-b919-4e7b-b2b2-4801fe56643b" (UID: "fac7e10c-b919-4e7b-b2b2-4801fe56643b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.251085 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4" (OuterVolumeSpecName: "kube-api-access-s9lt4") pod "fac7e10c-b919-4e7b-b2b2-4801fe56643b" (UID: "fac7e10c-b919-4e7b-b2b2-4801fe56643b"). InnerVolumeSpecName "kube-api-access-s9lt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.347372 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9lt4\" (UniqueName: \"kubernetes.io/projected/fac7e10c-b919-4e7b-b2b2-4801fe56643b-kube-api-access-s9lt4\") on node \"crc\" DevicePath \"\"" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.347417 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fac7e10c-b919-4e7b-b2b2-4801fe56643b-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.347429 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fac7e10c-b919-4e7b-b2b2-4801fe56643b-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.944949 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" event={"ID":"fac7e10c-b919-4e7b-b2b2-4801fe56643b","Type":"ContainerDied","Data":"33b6aa94609adcb06d33108015a2b18f2af0c66215e9977186cf847890582a1f"} Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.944995 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx" Sep 29 10:30:02 crc kubenswrapper[4727]: I0929 10:30:02.945002 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b6aa94609adcb06d33108015a2b18f2af0c66215e9977186cf847890582a1f" Sep 29 10:31:19 crc kubenswrapper[4727]: I0929 10:31:19.246593 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:31:19 crc kubenswrapper[4727]: I0929 10:31:19.247536 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:31:49 crc kubenswrapper[4727]: I0929 10:31:49.246624 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:31:49 crc kubenswrapper[4727]: I0929 10:31:49.247103 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.246811 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.247475 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.247535 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.248169 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.248219 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16" gracePeriod=600 Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.766980 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16" exitCode=0 Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.767116 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16"} Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.767338 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090"} Sep 29 10:32:19 crc kubenswrapper[4727]: I0929 10:32:19.767384 4727 scope.go:117] "RemoveContainer" containerID="5d6c6d418c810ac14745cea88746be23c10003dc31d856be14053ca2a6c3bdbc" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.725425 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-62b6k"] Sep 29 10:33:51 crc kubenswrapper[4727]: E0929 10:33:51.726423 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7e10c-b919-4e7b-b2b2-4801fe56643b" containerName="collect-profiles" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.726439 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7e10c-b919-4e7b-b2b2-4801fe56643b" containerName="collect-profiles" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.726549 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7e10c-b919-4e7b-b2b2-4801fe56643b" containerName="collect-profiles" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.727051 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.730240 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.733673 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.733801 4727 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b2vd4" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.746511 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wxxch"] Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.747671 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wxxch" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.749144 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-62b6k"] Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.749921 4727 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-gn5bf" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.756708 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6w54j"] Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.757604 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.762635 4727 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-lggz9" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.767688 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wxxch"] Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.783830 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8l4n\" (UniqueName: \"kubernetes.io/projected/4e4ebfd5-7776-4705-943f-b6ac0e6a6238-kube-api-access-k8l4n\") pod \"cert-manager-webhook-5655c58dd6-6w54j\" (UID: \"4e4ebfd5-7776-4705-943f-b6ac0e6a6238\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.783885 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5clt5\" (UniqueName: \"kubernetes.io/projected/15e48513-d028-479d-9796-ab2ecbe72f58-kube-api-access-5clt5\") pod \"cert-manager-cainjector-7f985d654d-62b6k\" (UID: \"15e48513-d028-479d-9796-ab2ecbe72f58\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.783930 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4gqb\" (UniqueName: \"kubernetes.io/projected/aec4835b-f46f-4b6a-aab2-37154553a6aa-kube-api-access-c4gqb\") pod \"cert-manager-5b446d88c5-wxxch\" (UID: \"aec4835b-f46f-4b6a-aab2-37154553a6aa\") " pod="cert-manager/cert-manager-5b446d88c5-wxxch" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.785559 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6w54j"] Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.884867 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5clt5\" (UniqueName: \"kubernetes.io/projected/15e48513-d028-479d-9796-ab2ecbe72f58-kube-api-access-5clt5\") pod \"cert-manager-cainjector-7f985d654d-62b6k\" (UID: \"15e48513-d028-479d-9796-ab2ecbe72f58\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.884934 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4gqb\" (UniqueName: \"kubernetes.io/projected/aec4835b-f46f-4b6a-aab2-37154553a6aa-kube-api-access-c4gqb\") pod \"cert-manager-5b446d88c5-wxxch\" (UID: \"aec4835b-f46f-4b6a-aab2-37154553a6aa\") " pod="cert-manager/cert-manager-5b446d88c5-wxxch" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.884974 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8l4n\" (UniqueName: \"kubernetes.io/projected/4e4ebfd5-7776-4705-943f-b6ac0e6a6238-kube-api-access-k8l4n\") pod \"cert-manager-webhook-5655c58dd6-6w54j\" (UID: \"4e4ebfd5-7776-4705-943f-b6ac0e6a6238\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.909096 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5clt5\" (UniqueName: \"kubernetes.io/projected/15e48513-d028-479d-9796-ab2ecbe72f58-kube-api-access-5clt5\") pod \"cert-manager-cainjector-7f985d654d-62b6k\" (UID: \"15e48513-d028-479d-9796-ab2ecbe72f58\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.909199 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8l4n\" (UniqueName: \"kubernetes.io/projected/4e4ebfd5-7776-4705-943f-b6ac0e6a6238-kube-api-access-k8l4n\") pod \"cert-manager-webhook-5655c58dd6-6w54j\" (UID: \"4e4ebfd5-7776-4705-943f-b6ac0e6a6238\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:51 crc kubenswrapper[4727]: I0929 10:33:51.909322 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4gqb\" (UniqueName: \"kubernetes.io/projected/aec4835b-f46f-4b6a-aab2-37154553a6aa-kube-api-access-c4gqb\") pod \"cert-manager-5b446d88c5-wxxch\" (UID: \"aec4835b-f46f-4b6a-aab2-37154553a6aa\") " pod="cert-manager/cert-manager-5b446d88c5-wxxch" Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.044060 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.073590 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wxxch" Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.083655 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.547841 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-62b6k"] Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.549917 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.553918 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wxxch"] Sep 29 10:33:52 crc kubenswrapper[4727]: W0929 10:33:52.554178 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaec4835b_f46f_4b6a_aab2_37154553a6aa.slice/crio-09d7c3f0438d8859b02953f8caf4da173bbbc154cbc047cd74877d16af68d8ab WatchSource:0}: Error finding container 09d7c3f0438d8859b02953f8caf4da173bbbc154cbc047cd74877d16af68d8ab: Status 404 returned error can't find the container with id 09d7c3f0438d8859b02953f8caf4da173bbbc154cbc047cd74877d16af68d8ab Sep 29 10:33:52 crc kubenswrapper[4727]: I0929 10:33:52.558184 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6w54j"] Sep 29 10:33:53 crc kubenswrapper[4727]: I0929 10:33:53.264090 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" event={"ID":"4e4ebfd5-7776-4705-943f-b6ac0e6a6238","Type":"ContainerStarted","Data":"3e4ff28d9e5c5792536e3ee8bd1a23d9282f05df8b2d19b457d9b768b71abd43"} Sep 29 10:33:53 crc kubenswrapper[4727]: I0929 10:33:53.265267 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wxxch" event={"ID":"aec4835b-f46f-4b6a-aab2-37154553a6aa","Type":"ContainerStarted","Data":"09d7c3f0438d8859b02953f8caf4da173bbbc154cbc047cd74877d16af68d8ab"} Sep 29 10:33:53 crc kubenswrapper[4727]: I0929 10:33:53.266757 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" event={"ID":"15e48513-d028-479d-9796-ab2ecbe72f58","Type":"ContainerStarted","Data":"8d32a0d35b4a8ef74b9cfd56b4775c31cd06f2214173aa5ad5d770a8911177da"} Sep 29 10:33:56 crc kubenswrapper[4727]: I0929 10:33:56.289086 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" event={"ID":"15e48513-d028-479d-9796-ab2ecbe72f58","Type":"ContainerStarted","Data":"14fb5da65849c3b74bdb4f49106ad7449f73268314b4b72ad9314e771df52202"} Sep 29 10:33:56 crc kubenswrapper[4727]: I0929 10:33:56.291176 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" event={"ID":"4e4ebfd5-7776-4705-943f-b6ac0e6a6238","Type":"ContainerStarted","Data":"6c15acfeb65194a2c304cfa6b4648a173973d26d6794688e38bf396f8af516ac"} Sep 29 10:33:56 crc kubenswrapper[4727]: I0929 10:33:56.291651 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:33:56 crc kubenswrapper[4727]: I0929 10:33:56.315922 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-62b6k" podStartSLOduration=1.971974367 podStartE2EDuration="5.315886646s" podCreationTimestamp="2025-09-29 10:33:51 +0000 UTC" firstStartedPulling="2025-09-29 10:33:52.549622118 +0000 UTC m=+702.722935480" lastFinishedPulling="2025-09-29 10:33:55.893534387 +0000 UTC m=+706.066847759" observedRunningTime="2025-09-29 10:33:56.311883432 +0000 UTC m=+706.485196794" watchObservedRunningTime="2025-09-29 10:33:56.315886646 +0000 UTC m=+706.489200018" Sep 29 10:33:56 crc kubenswrapper[4727]: I0929 10:33:56.342558 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" podStartSLOduration=1.96822426 podStartE2EDuration="5.342530605s" podCreationTimestamp="2025-09-29 10:33:51 +0000 UTC" firstStartedPulling="2025-09-29 10:33:52.565011016 +0000 UTC m=+702.738324378" lastFinishedPulling="2025-09-29 10:33:55.939317361 +0000 UTC m=+706.112630723" observedRunningTime="2025-09-29 10:33:56.339108896 +0000 UTC m=+706.512422258" watchObservedRunningTime="2025-09-29 10:33:56.342530605 +0000 UTC m=+706.515843967" Sep 29 10:33:58 crc kubenswrapper[4727]: I0929 10:33:58.306594 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wxxch" event={"ID":"aec4835b-f46f-4b6a-aab2-37154553a6aa","Type":"ContainerStarted","Data":"c5bd73bbc844fa6ead9fc9a8c56a848eaf3c2bea649dbb82c3139041e91bb840"} Sep 29 10:33:58 crc kubenswrapper[4727]: I0929 10:33:58.335112 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-wxxch" podStartSLOduration=2.446069713 podStartE2EDuration="7.335084017s" podCreationTimestamp="2025-09-29 10:33:51 +0000 UTC" firstStartedPulling="2025-09-29 10:33:52.556496816 +0000 UTC m=+702.729810178" lastFinishedPulling="2025-09-29 10:33:57.44551112 +0000 UTC m=+707.618824482" observedRunningTime="2025-09-29 10:33:58.334385259 +0000 UTC m=+708.507698621" watchObservedRunningTime="2025-09-29 10:33:58.335084017 +0000 UTC m=+708.508397379" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.006002 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lqql2"] Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.006764 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-controller" containerID="cri-o://41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.006941 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="northd" containerID="cri-o://de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.006982 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.007010 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-node" containerID="cri-o://839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.007039 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-acl-logging" containerID="cri-o://f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.007072 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="sbdb" containerID="cri-o://e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.006802 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="nbdb" containerID="cri-o://01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.050156 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" containerID="cri-o://7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" gracePeriod=30 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.087452 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-6w54j" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.342061 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovnkube-controller/3.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.347822 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-acl-logging/0.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.349563 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-controller/0.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350083 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" exitCode=0 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350109 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" exitCode=0 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350118 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" exitCode=0 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350155 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" exitCode=143 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350165 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" exitCode=143 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350244 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350326 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350377 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350399 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350404 4727 scope.go:117] "RemoveContainer" containerID="66f3f5ee052331f3f748ece21abe82eeae2f650717cce75dd9e5da6d1dca344e" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.350419 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.353187 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/2.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.353726 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/1.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.353772 4727 generic.go:334] "Generic (PLEG): container finished" podID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" containerID="0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa" exitCode=2 Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.353805 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerDied","Data":"0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa"} Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.354509 4727 scope.go:117] "RemoveContainer" containerID="0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.354710 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xwdcz_openshift-multus(9e3b98d9-a9fc-4bf3-8053-b7701c047d99)\"" pod="openshift-multus/multus-xwdcz" podUID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.439510 4727 scope.go:117] "RemoveContainer" containerID="e5007e1107f2b9c31e59ccc185ce77b4a0584c6d00947357001d801aca43a39f" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.876986 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-acl-logging/0.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.877666 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-controller/0.log" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.878387 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.941819 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jxg57"] Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.942991 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="sbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943065 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="sbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943097 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="nbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943114 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="nbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943134 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="northd" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943257 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="northd" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943285 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943307 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943331 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943376 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943393 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943405 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943423 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943437 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943452 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943466 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943486 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-acl-logging" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943499 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-acl-logging" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943518 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-node" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943531 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-node" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943551 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943564 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.943585 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kubecfg-setup" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943598 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kubecfg-setup" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943812 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943832 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="northd" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943856 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="nbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943881 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943900 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-acl-logging" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943916 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943941 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="sbdb" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.943974 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.944002 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="kube-rbac-proxy-node" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.944025 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovn-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: E0929 10:34:02.944522 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.944544 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.944764 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.945182 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerName="ovnkube-controller" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.950035 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.953805 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.953855 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq6pj\" (UniqueName: \"kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.953908 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.953929 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.953984 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954005 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954048 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954067 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954091 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954161 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954215 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954271 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954308 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954328 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954378 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954410 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954435 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954464 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954462 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log" (OuterVolumeSpecName: "node-log") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954539 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954542 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954583 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954484 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954661 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units\") pod \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\" (UID: \"cc578ace-78c8-4d17-a556-c6d6ceb149a4\") " Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955060 4727 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-node-log\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955080 4727 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955092 4727 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955102 4727 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954610 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954634 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954683 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954663 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954717 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954733 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash" (OuterVolumeSpecName: "host-slash") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954768 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954799 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.954830 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket" (OuterVolumeSpecName: "log-socket") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955046 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955096 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955095 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.955896 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.962254 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.965612 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj" (OuterVolumeSpecName: "kube-api-access-qq6pj") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "kube-api-access-qq6pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:34:02 crc kubenswrapper[4727]: I0929 10:34:02.970372 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "cc578ace-78c8-4d17-a556-c6d6ceb149a4" (UID: "cc578ace-78c8-4d17-a556-c6d6ceb149a4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.056530 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-node-log\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.056932 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.056980 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-slash\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057015 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63850845-5110-4649-a3ae-9f8d462a9d48-ovn-node-metrics-cert\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057045 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-netns\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057069 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-kubelet\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057105 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-config\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057131 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-var-lib-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057165 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057190 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057360 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-systemd-units\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057391 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-ovn\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057412 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmntq\" (UniqueName: \"kubernetes.io/projected/63850845-5110-4649-a3ae-9f8d462a9d48-kube-api-access-dmntq\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057435 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-netd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057457 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-log-socket\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057478 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-bin\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057505 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-script-lib\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057527 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-etc-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057551 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-env-overrides\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057583 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-systemd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057633 4727 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057647 4727 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057659 4727 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057671 4727 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057686 4727 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057697 4727 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057708 4727 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057718 4727 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-log-socket\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057729 4727 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057741 4727 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-slash\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057752 4727 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057762 4727 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057773 4727 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057785 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq6pj\" (UniqueName: \"kubernetes.io/projected/cc578ace-78c8-4d17-a556-c6d6ceb149a4-kube-api-access-qq6pj\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057797 4727 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cc578ace-78c8-4d17-a556-c6d6ceb149a4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.057810 4727 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc578ace-78c8-4d17-a556-c6d6ceb149a4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.158917 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-systemd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159012 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-node-log\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159109 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-node-log\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159043 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159168 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-slash\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159181 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159224 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-slash\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159248 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63850845-5110-4649-a3ae-9f8d462a9d48-ovn-node-metrics-cert\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159272 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-netns\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159285 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-systemd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159313 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-kubelet\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159361 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-var-lib-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159366 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-netns\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159384 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-config\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159398 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-var-lib-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159440 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-kubelet\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159507 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159553 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159622 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-run-ovn-kubernetes\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159639 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159694 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-systemd-units\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159715 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-ovn\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159721 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-systemd-units\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159732 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmntq\" (UniqueName: \"kubernetes.io/projected/63850845-5110-4649-a3ae-9f8d462a9d48-kube-api-access-dmntq\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-netd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159760 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-run-ovn\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159773 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-log-socket\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159791 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-bin\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159815 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-script-lib\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159835 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-etc-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.159859 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-env-overrides\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160047 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-etc-openvswitch\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160086 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-bin\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-host-cni-netd\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160171 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/63850845-5110-4649-a3ae-9f8d462a9d48-log-socket\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160312 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-config\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160631 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-ovnkube-script-lib\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.160744 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63850845-5110-4649-a3ae-9f8d462a9d48-env-overrides\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.163947 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63850845-5110-4649-a3ae-9f8d462a9d48-ovn-node-metrics-cert\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.176452 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmntq\" (UniqueName: \"kubernetes.io/projected/63850845-5110-4649-a3ae-9f8d462a9d48-kube-api-access-dmntq\") pod \"ovnkube-node-jxg57\" (UID: \"63850845-5110-4649-a3ae-9f8d462a9d48\") " pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.269566 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.359144 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"2d20950b4c28eeb131ecdfe78d9d269e1f6dd2032dfe667c23da89127c981a66"} Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.361905 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/2.log" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.367659 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-acl-logging/0.log" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.368377 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqql2_cc578ace-78c8-4d17-a556-c6d6ceb149a4/ovn-controller/0.log" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369036 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" exitCode=0 Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369086 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" exitCode=0 Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369096 4727 generic.go:334] "Generic (PLEG): container finished" podID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" exitCode=0 Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369159 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795"} Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369218 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66"} Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369239 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625"} Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369256 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" event={"ID":"cc578ace-78c8-4d17-a556-c6d6ceb149a4","Type":"ContainerDied","Data":"c69c1240554da2879557b426f455d8e7615dff9d27e752bdc81466afa2e60df8"} Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369303 4727 scope.go:117] "RemoveContainer" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.369578 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqql2" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.402504 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lqql2"] Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.405635 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lqql2"] Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.417553 4727 scope.go:117] "RemoveContainer" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.435815 4727 scope.go:117] "RemoveContainer" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.449124 4727 scope.go:117] "RemoveContainer" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.468559 4727 scope.go:117] "RemoveContainer" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.487422 4727 scope.go:117] "RemoveContainer" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.507721 4727 scope.go:117] "RemoveContainer" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.524323 4727 scope.go:117] "RemoveContainer" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.543875 4727 scope.go:117] "RemoveContainer" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.605669 4727 scope.go:117] "RemoveContainer" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.606404 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": container with ID starting with 7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce not found: ID does not exist" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.606450 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce"} err="failed to get container status \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": rpc error: code = NotFound desc = could not find container \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": container with ID starting with 7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.606483 4727 scope.go:117] "RemoveContainer" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.606894 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": container with ID starting with e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795 not found: ID does not exist" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.606949 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795"} err="failed to get container status \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": rpc error: code = NotFound desc = could not find container \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": container with ID starting with e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.606994 4727 scope.go:117] "RemoveContainer" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.607530 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": container with ID starting with 01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66 not found: ID does not exist" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.607596 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66"} err="failed to get container status \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": rpc error: code = NotFound desc = could not find container \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": container with ID starting with 01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.607642 4727 scope.go:117] "RemoveContainer" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.608178 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": container with ID starting with de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625 not found: ID does not exist" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.608215 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625"} err="failed to get container status \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": rpc error: code = NotFound desc = could not find container \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": container with ID starting with de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.608239 4727 scope.go:117] "RemoveContainer" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.608687 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": container with ID starting with 28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb not found: ID does not exist" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.608782 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb"} err="failed to get container status \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": rpc error: code = NotFound desc = could not find container \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": container with ID starting with 28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.608853 4727 scope.go:117] "RemoveContainer" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.609362 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": container with ID starting with 839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373 not found: ID does not exist" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.609394 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373"} err="failed to get container status \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": rpc error: code = NotFound desc = could not find container \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": container with ID starting with 839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.609416 4727 scope.go:117] "RemoveContainer" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.610007 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": container with ID starting with f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943 not found: ID does not exist" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.610047 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943"} err="failed to get container status \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": rpc error: code = NotFound desc = could not find container \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": container with ID starting with f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.610069 4727 scope.go:117] "RemoveContainer" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.610485 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": container with ID starting with 41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f not found: ID does not exist" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.610546 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f"} err="failed to get container status \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": rpc error: code = NotFound desc = could not find container \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": container with ID starting with 41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.610566 4727 scope.go:117] "RemoveContainer" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" Sep 29 10:34:03 crc kubenswrapper[4727]: E0929 10:34:03.610959 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": container with ID starting with ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983 not found: ID does not exist" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.610998 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983"} err="failed to get container status \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": rpc error: code = NotFound desc = could not find container \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": container with ID starting with ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.611018 4727 scope.go:117] "RemoveContainer" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.611403 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce"} err="failed to get container status \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": rpc error: code = NotFound desc = could not find container \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": container with ID starting with 7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.611429 4727 scope.go:117] "RemoveContainer" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.611814 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795"} err="failed to get container status \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": rpc error: code = NotFound desc = could not find container \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": container with ID starting with e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.611839 4727 scope.go:117] "RemoveContainer" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.612412 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66"} err="failed to get container status \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": rpc error: code = NotFound desc = could not find container \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": container with ID starting with 01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.612442 4727 scope.go:117] "RemoveContainer" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.612682 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625"} err="failed to get container status \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": rpc error: code = NotFound desc = could not find container \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": container with ID starting with de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.612702 4727 scope.go:117] "RemoveContainer" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.612978 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb"} err="failed to get container status \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": rpc error: code = NotFound desc = could not find container \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": container with ID starting with 28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.613009 4727 scope.go:117] "RemoveContainer" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.613377 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373"} err="failed to get container status \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": rpc error: code = NotFound desc = could not find container \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": container with ID starting with 839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.613402 4727 scope.go:117] "RemoveContainer" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.613712 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943"} err="failed to get container status \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": rpc error: code = NotFound desc = could not find container \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": container with ID starting with f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.613737 4727 scope.go:117] "RemoveContainer" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.614092 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f"} err="failed to get container status \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": rpc error: code = NotFound desc = could not find container \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": container with ID starting with 41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.614124 4727 scope.go:117] "RemoveContainer" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.614718 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983"} err="failed to get container status \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": rpc error: code = NotFound desc = could not find container \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": container with ID starting with ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.614749 4727 scope.go:117] "RemoveContainer" containerID="7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615033 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce"} err="failed to get container status \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": rpc error: code = NotFound desc = could not find container \"7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce\": container with ID starting with 7374f5d2155ce309e71712fb0b5a95d5ee0e7f472688a8820a8fad4fe14cecce not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615056 4727 scope.go:117] "RemoveContainer" containerID="e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615470 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795"} err="failed to get container status \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": rpc error: code = NotFound desc = could not find container \"e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795\": container with ID starting with e1c7a418e04639f77338950243e4ead88f6829c51da5159d7d42243bc9878795 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615535 4727 scope.go:117] "RemoveContainer" containerID="01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615891 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66"} err="failed to get container status \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": rpc error: code = NotFound desc = could not find container \"01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66\": container with ID starting with 01d84e68eaa7e929480fa22458aec9539d0dd14a9a938d8f4831f0beb109ae66 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.615915 4727 scope.go:117] "RemoveContainer" containerID="de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.616329 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625"} err="failed to get container status \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": rpc error: code = NotFound desc = could not find container \"de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625\": container with ID starting with de9696388b6409459cbdd5fa29bdd8fa11e6b959604a7e9c4785850827a10625 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.616379 4727 scope.go:117] "RemoveContainer" containerID="28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.616815 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb"} err="failed to get container status \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": rpc error: code = NotFound desc = could not find container \"28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb\": container with ID starting with 28c138df8a84f729e2cf4bd6654cab8eb2961ed960a8dfda3013040a555c10cb not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.616865 4727 scope.go:117] "RemoveContainer" containerID="839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.617209 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373"} err="failed to get container status \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": rpc error: code = NotFound desc = could not find container \"839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373\": container with ID starting with 839504f36d5605c1cfbfe4b59ee12bba5deec3438c3ef46b725df9d15e291373 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.617242 4727 scope.go:117] "RemoveContainer" containerID="f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.617583 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943"} err="failed to get container status \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": rpc error: code = NotFound desc = could not find container \"f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943\": container with ID starting with f151c5627ee02d403ea0ec904bb03bd190c7836b2c63fc501cf520e0903a6943 not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.617611 4727 scope.go:117] "RemoveContainer" containerID="41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.618018 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f"} err="failed to get container status \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": rpc error: code = NotFound desc = could not find container \"41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f\": container with ID starting with 41a82d3704df19660758c08e204e2e46c715cc02dee20c453616c9156e04db8f not found: ID does not exist" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.618053 4727 scope.go:117] "RemoveContainer" containerID="ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983" Sep 29 10:34:03 crc kubenswrapper[4727]: I0929 10:34:03.618464 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983"} err="failed to get container status \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": rpc error: code = NotFound desc = could not find container \"ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983\": container with ID starting with ca0cfac6609eb27e99ebd5f3a15ef596e2a8f6fc0ff35194f52e6b1e940b1983 not found: ID does not exist" Sep 29 10:34:04 crc kubenswrapper[4727]: I0929 10:34:04.376077 4727 generic.go:334] "Generic (PLEG): container finished" podID="63850845-5110-4649-a3ae-9f8d462a9d48" containerID="57e3e09eaef6a32c5c42ae157b8d79974c6cdbd16a23aebd7b1ea442aab31d2d" exitCode=0 Sep 29 10:34:04 crc kubenswrapper[4727]: I0929 10:34:04.376157 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerDied","Data":"57e3e09eaef6a32c5c42ae157b8d79974c6cdbd16a23aebd7b1ea442aab31d2d"} Sep 29 10:34:05 crc kubenswrapper[4727]: I0929 10:34:05.116134 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc578ace-78c8-4d17-a556-c6d6ceb149a4" path="/var/lib/kubelet/pods/cc578ace-78c8-4d17-a556-c6d6ceb149a4/volumes" Sep 29 10:34:05 crc kubenswrapper[4727]: I0929 10:34:05.389555 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"bebefb0b1e98d04f406b0915a15050f82348921ce2c3f7dcab14b39993d5febe"} Sep 29 10:34:05 crc kubenswrapper[4727]: I0929 10:34:05.389614 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"736c7ccafd0b367159a45ed29a8de5cafd28bce494f2df0c1302d76f56390530"} Sep 29 10:34:05 crc kubenswrapper[4727]: I0929 10:34:05.389627 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"6463e3409b55e207fda13c1f9bd854e5f61e00cc9556d9fe0f8a7f52324776ef"} Sep 29 10:34:06 crc kubenswrapper[4727]: I0929 10:34:06.400445 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"a0dc2cad96326a30d5ce91137e383d64f3807b018cf7b852bc8958c65580d624"} Sep 29 10:34:06 crc kubenswrapper[4727]: I0929 10:34:06.400777 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"bf58ac84668a1e599311d564e60a924908bcd663e5dbd25d401706d4d753a473"} Sep 29 10:34:06 crc kubenswrapper[4727]: I0929 10:34:06.400788 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"be09b49940c77208f139af63251c77250cf9e994ed87d8e8810acdfaff9cbd44"} Sep 29 10:34:08 crc kubenswrapper[4727]: I0929 10:34:08.416935 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"26320ad549085c98c996fd2e3cbcf25b00bd99d99d54850ac2702e4e5603fb02"} Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.436910 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" event={"ID":"63850845-5110-4649-a3ae-9f8d462a9d48","Type":"ContainerStarted","Data":"5827755bb2fc9fcbe0773885a70b6b9fe399d33df2683d35245adfa4471e0760"} Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.438097 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.438145 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.438157 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.497175 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.499125 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:11 crc kubenswrapper[4727]: I0929 10:34:11.510842 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" podStartSLOduration=9.510818533 podStartE2EDuration="9.510818533s" podCreationTimestamp="2025-09-29 10:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:34:11.504626544 +0000 UTC m=+721.677939906" watchObservedRunningTime="2025-09-29 10:34:11.510818533 +0000 UTC m=+721.684131895" Sep 29 10:34:14 crc kubenswrapper[4727]: I0929 10:34:14.107701 4727 scope.go:117] "RemoveContainer" containerID="0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa" Sep 29 10:34:14 crc kubenswrapper[4727]: E0929 10:34:14.108306 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xwdcz_openshift-multus(9e3b98d9-a9fc-4bf3-8053-b7701c047d99)\"" pod="openshift-multus/multus-xwdcz" podUID="9e3b98d9-a9fc-4bf3-8053-b7701c047d99" Sep 29 10:34:19 crc kubenswrapper[4727]: I0929 10:34:19.246330 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:34:19 crc kubenswrapper[4727]: I0929 10:34:19.247042 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:34:26 crc kubenswrapper[4727]: I0929 10:34:26.108627 4727 scope.go:117] "RemoveContainer" containerID="0f90ace1d97690fe3295192c2d6cc108abf8fae33f5cbd1ca527d3df57ea75aa" Sep 29 10:34:26 crc kubenswrapper[4727]: I0929 10:34:26.518611 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwdcz_9e3b98d9-a9fc-4bf3-8053-b7701c047d99/kube-multus/2.log" Sep 29 10:34:26 crc kubenswrapper[4727]: I0929 10:34:26.519268 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwdcz" event={"ID":"9e3b98d9-a9fc-4bf3-8053-b7701c047d99","Type":"ContainerStarted","Data":"7ef7f1cefe2500e73fd8d24a4f9c87c0a8efce9c1c8d16a0c0f22811947121e3"} Sep 29 10:34:33 crc kubenswrapper[4727]: I0929 10:34:33.322456 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jxg57" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.187133 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk"] Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.188677 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.190646 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.197404 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk"] Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.351275 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7b4\" (UniqueName: \"kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.351697 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.351739 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.452922 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.453007 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7b4\" (UniqueName: \"kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.453043 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.453427 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.453437 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.480746 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7b4\" (UniqueName: \"kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.507950 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:42 crc kubenswrapper[4727]: I0929 10:34:42.687680 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk"] Sep 29 10:34:43 crc kubenswrapper[4727]: I0929 10:34:43.607299 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerStarted","Data":"bbdf735aa5e66fb6f180aff3f11c54627af922966a6693a3e8e13dc93bb13717"} Sep 29 10:34:43 crc kubenswrapper[4727]: I0929 10:34:43.607641 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerStarted","Data":"599dc056d1443753c80265d7fa5886acbf1710f9f3040d41451ff1ddd2721e82"} Sep 29 10:34:44 crc kubenswrapper[4727]: I0929 10:34:44.613833 4727 generic.go:334] "Generic (PLEG): container finished" podID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerID="bbdf735aa5e66fb6f180aff3f11c54627af922966a6693a3e8e13dc93bb13717" exitCode=0 Sep 29 10:34:44 crc kubenswrapper[4727]: I0929 10:34:44.613890 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerDied","Data":"bbdf735aa5e66fb6f180aff3f11c54627af922966a6693a3e8e13dc93bb13717"} Sep 29 10:34:46 crc kubenswrapper[4727]: I0929 10:34:46.625233 4727 generic.go:334] "Generic (PLEG): container finished" podID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerID="1b8525cdce9a88ed7f0b3622ef3b001e5690355e0e3abb4d5b63db44c241bc7d" exitCode=0 Sep 29 10:34:46 crc kubenswrapper[4727]: I0929 10:34:46.625286 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerDied","Data":"1b8525cdce9a88ed7f0b3622ef3b001e5690355e0e3abb4d5b63db44c241bc7d"} Sep 29 10:34:47 crc kubenswrapper[4727]: I0929 10:34:47.635095 4727 generic.go:334] "Generic (PLEG): container finished" podID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerID="cad071314a7214c57e53c40d6555e9045d7e2e1ddf66649c0bf31fc37d9d7c22" exitCode=0 Sep 29 10:34:47 crc kubenswrapper[4727]: I0929 10:34:47.635191 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerDied","Data":"cad071314a7214c57e53c40d6555e9045d7e2e1ddf66649c0bf31fc37d9d7c22"} Sep 29 10:34:48 crc kubenswrapper[4727]: I0929 10:34:48.861253 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.042768 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl7b4\" (UniqueName: \"kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4\") pod \"2ae491e2-d46d-44be-9b32-877cc62c18e4\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.042827 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util\") pod \"2ae491e2-d46d-44be-9b32-877cc62c18e4\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.042867 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle\") pod \"2ae491e2-d46d-44be-9b32-877cc62c18e4\" (UID: \"2ae491e2-d46d-44be-9b32-877cc62c18e4\") " Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.043593 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle" (OuterVolumeSpecName: "bundle") pod "2ae491e2-d46d-44be-9b32-877cc62c18e4" (UID: "2ae491e2-d46d-44be-9b32-877cc62c18e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.049863 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4" (OuterVolumeSpecName: "kube-api-access-pl7b4") pod "2ae491e2-d46d-44be-9b32-877cc62c18e4" (UID: "2ae491e2-d46d-44be-9b32-877cc62c18e4"). InnerVolumeSpecName "kube-api-access-pl7b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.053407 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util" (OuterVolumeSpecName: "util") pod "2ae491e2-d46d-44be-9b32-877cc62c18e4" (UID: "2ae491e2-d46d-44be-9b32-877cc62c18e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.143832 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl7b4\" (UniqueName: \"kubernetes.io/projected/2ae491e2-d46d-44be-9b32-877cc62c18e4-kube-api-access-pl7b4\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.143879 4727 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-util\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.143892 4727 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ae491e2-d46d-44be-9b32-877cc62c18e4-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.246213 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.246274 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.647183 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" event={"ID":"2ae491e2-d46d-44be-9b32-877cc62c18e4","Type":"ContainerDied","Data":"599dc056d1443753c80265d7fa5886acbf1710f9f3040d41451ff1ddd2721e82"} Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.647225 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="599dc056d1443753c80265d7fa5886acbf1710f9f3040d41451ff1ddd2721e82" Sep 29 10:34:49 crc kubenswrapper[4727]: I0929 10:34:49.647230 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.646974 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth"] Sep 29 10:34:50 crc kubenswrapper[4727]: E0929 10:34:50.647568 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="extract" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.647580 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="extract" Sep 29 10:34:50 crc kubenswrapper[4727]: E0929 10:34:50.647592 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="pull" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.647598 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="pull" Sep 29 10:34:50 crc kubenswrapper[4727]: E0929 10:34:50.647610 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="util" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.647617 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="util" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.647728 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae491e2-d46d-44be-9b32-877cc62c18e4" containerName="extract" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.648121 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.651275 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.654108 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.654590 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xp486" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.697136 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth"] Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.761600 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnmjb\" (UniqueName: \"kubernetes.io/projected/2ef4df09-2078-4ef8-9c7b-8e637b20cfcf-kube-api-access-mnmjb\") pod \"nmstate-operator-5d6f6cfd66-5vcth\" (UID: \"2ef4df09-2078-4ef8-9c7b-8e637b20cfcf\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.862743 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnmjb\" (UniqueName: \"kubernetes.io/projected/2ef4df09-2078-4ef8-9c7b-8e637b20cfcf-kube-api-access-mnmjb\") pod \"nmstate-operator-5d6f6cfd66-5vcth\" (UID: \"2ef4df09-2078-4ef8-9c7b-8e637b20cfcf\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.887301 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnmjb\" (UniqueName: \"kubernetes.io/projected/2ef4df09-2078-4ef8-9c7b-8e637b20cfcf-kube-api-access-mnmjb\") pod \"nmstate-operator-5d6f6cfd66-5vcth\" (UID: \"2ef4df09-2078-4ef8-9c7b-8e637b20cfcf\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" Sep 29 10:34:50 crc kubenswrapper[4727]: I0929 10:34:50.963618 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" Sep 29 10:34:51 crc kubenswrapper[4727]: I0929 10:34:51.138910 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth"] Sep 29 10:34:51 crc kubenswrapper[4727]: W0929 10:34:51.150850 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ef4df09_2078_4ef8_9c7b_8e637b20cfcf.slice/crio-5f71e143d52da8dfc932f7bfbad9b039bf516015e70e4dfd192bb4ad85552544 WatchSource:0}: Error finding container 5f71e143d52da8dfc932f7bfbad9b039bf516015e70e4dfd192bb4ad85552544: Status 404 returned error can't find the container with id 5f71e143d52da8dfc932f7bfbad9b039bf516015e70e4dfd192bb4ad85552544 Sep 29 10:34:51 crc kubenswrapper[4727]: I0929 10:34:51.658691 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" event={"ID":"2ef4df09-2078-4ef8-9c7b-8e637b20cfcf","Type":"ContainerStarted","Data":"5f71e143d52da8dfc932f7bfbad9b039bf516015e70e4dfd192bb4ad85552544"} Sep 29 10:34:54 crc kubenswrapper[4727]: I0929 10:34:54.674791 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" event={"ID":"2ef4df09-2078-4ef8-9c7b-8e637b20cfcf","Type":"ContainerStarted","Data":"3b66bfa8cfc52517e321d84405cd2861f65b263d3701033059ea4673bc346bfe"} Sep 29 10:34:54 crc kubenswrapper[4727]: I0929 10:34:54.691901 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-5vcth" podStartSLOduration=2.352959554 podStartE2EDuration="4.69188563s" podCreationTimestamp="2025-09-29 10:34:50 +0000 UTC" firstStartedPulling="2025-09-29 10:34:51.153722648 +0000 UTC m=+761.327036010" lastFinishedPulling="2025-09-29 10:34:53.492648724 +0000 UTC m=+763.665962086" observedRunningTime="2025-09-29 10:34:54.688624956 +0000 UTC m=+764.861938318" watchObservedRunningTime="2025-09-29 10:34:54.69188563 +0000 UTC m=+764.865198992" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.707961 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2hx48"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.709406 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.715178 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-mbmms" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.718542 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-shptq"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.719227 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.728143 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.750428 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2hx48"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.754580 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-shptq"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.764450 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-sh4hc"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.765388 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.818929 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.818999 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5f4\" (UniqueName: \"kubernetes.io/projected/7655e999-03ae-45f0-a123-1efa69626f7b-kube-api-access-5n5f4\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.819053 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7bg7\" (UniqueName: \"kubernetes.io/projected/a713c14e-bce4-44a3-940d-0a987cc6fa5b-kube-api-access-d7bg7\") pod \"nmstate-metrics-58fcddf996-2hx48\" (UID: \"a713c14e-bce4-44a3-940d-0a987cc6fa5b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.901774 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv"] Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.902584 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.905609 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.905998 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.906002 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-wlh5p" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.919758 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-ovs-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.919799 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-nmstate-lock\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.919818 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfn48\" (UniqueName: \"kubernetes.io/projected/8109dd98-72c1-4542-a34c-1f5d911b5fd4-kube-api-access-hfn48\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.919843 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.920007 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5f4\" (UniqueName: \"kubernetes.io/projected/7655e999-03ae-45f0-a123-1efa69626f7b-kube-api-access-5n5f4\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.920107 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7bg7\" (UniqueName: \"kubernetes.io/projected/a713c14e-bce4-44a3-940d-0a987cc6fa5b-kube-api-access-d7bg7\") pod \"nmstate-metrics-58fcddf996-2hx48\" (UID: \"a713c14e-bce4-44a3-940d-0a987cc6fa5b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.920149 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-dbus-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:55 crc kubenswrapper[4727]: E0929 10:34:55.920013 4727 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Sep 29 10:34:55 crc kubenswrapper[4727]: E0929 10:34:55.920276 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair podName:7655e999-03ae-45f0-a123-1efa69626f7b nodeName:}" failed. No retries permitted until 2025-09-29 10:34:56.420257915 +0000 UTC m=+766.593571277 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair") pod "nmstate-webhook-6d689559c5-shptq" (UID: "7655e999-03ae-45f0-a123-1efa69626f7b") : secret "openshift-nmstate-webhook" not found Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.944359 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7bg7\" (UniqueName: \"kubernetes.io/projected/a713c14e-bce4-44a3-940d-0a987cc6fa5b-kube-api-access-d7bg7\") pod \"nmstate-metrics-58fcddf996-2hx48\" (UID: \"a713c14e-bce4-44a3-940d-0a987cc6fa5b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.944397 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5f4\" (UniqueName: \"kubernetes.io/projected/7655e999-03ae-45f0-a123-1efa69626f7b-kube-api-access-5n5f4\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:55 crc kubenswrapper[4727]: I0929 10:34:55.959141 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv"] Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021410 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021470 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkrms\" (UniqueName: \"kubernetes.io/projected/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-kube-api-access-lkrms\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021523 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021545 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-dbus-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021579 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-ovs-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021644 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-ovs-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021656 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-nmstate-lock\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021681 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-nmstate-lock\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021703 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfn48\" (UniqueName: \"kubernetes.io/projected/8109dd98-72c1-4542-a34c-1f5d911b5fd4-kube-api-access-hfn48\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.021944 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8109dd98-72c1-4542-a34c-1f5d911b5fd4-dbus-socket\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.025961 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.044387 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfn48\" (UniqueName: \"kubernetes.io/projected/8109dd98-72c1-4542-a34c-1f5d911b5fd4-kube-api-access-hfn48\") pod \"nmstate-handler-sh4hc\" (UID: \"8109dd98-72c1-4542-a34c-1f5d911b5fd4\") " pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.087193 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:34:56 crc kubenswrapper[4727]: W0929 10:34:56.106921 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8109dd98_72c1_4542_a34c_1f5d911b5fd4.slice/crio-c51aca9e82f53c5d9b4a017dd3370e33563ee91bcb048e802736ad6f2fef1c56 WatchSource:0}: Error finding container c51aca9e82f53c5d9b4a017dd3370e33563ee91bcb048e802736ad6f2fef1c56: Status 404 returned error can't find the container with id c51aca9e82f53c5d9b4a017dd3370e33563ee91bcb048e802736ad6f2fef1c56 Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.123902 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.123995 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.124022 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkrms\" (UniqueName: \"kubernetes.io/projected/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-kube-api-access-lkrms\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.125635 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.128572 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.140193 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkrms\" (UniqueName: \"kubernetes.io/projected/8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6-kube-api-access-lkrms\") pod \"nmstate-console-plugin-864bb6dfb5-l7xjv\" (UID: \"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.216061 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.233897 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2hx48"] Sep 29 10:34:56 crc kubenswrapper[4727]: W0929 10:34:56.238105 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda713c14e_bce4_44a3_940d_0a987cc6fa5b.slice/crio-554ff638b0ca662fa94ce2fea761bb54bff225199522ec5612edd645c8c70202 WatchSource:0}: Error finding container 554ff638b0ca662fa94ce2fea761bb54bff225199522ec5612edd645c8c70202: Status 404 returned error can't find the container with id 554ff638b0ca662fa94ce2fea761bb54bff225199522ec5612edd645c8c70202 Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.366847 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7bb44447b-b8xnt"] Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.367994 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.387135 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7bb44447b-b8xnt"] Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.411230 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv"] Sep 29 10:34:56 crc kubenswrapper[4727]: W0929 10:34:56.417040 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a2bcf35_f31c_44b3_8a73_c6c6e2c5c2c6.slice/crio-abc8fcb4fd52beeb0ed3464f3453f0f83d670ffc59c8ea9593e9b81d6483af63 WatchSource:0}: Error finding container abc8fcb4fd52beeb0ed3464f3453f0f83d670ffc59c8ea9593e9b81d6483af63: Status 404 returned error can't find the container with id abc8fcb4fd52beeb0ed3464f3453f0f83d670ffc59c8ea9593e9b81d6483af63 Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.426891 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.431917 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7655e999-03ae-45f0-a123-1efa69626f7b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-shptq\" (UID: \"7655e999-03ae-45f0-a123-1efa69626f7b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.527967 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-service-ca\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528033 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528216 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-trusted-ca-bundle\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-oauth-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528409 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brlrh\" (UniqueName: \"kubernetes.io/projected/b5d6ff30-da65-4a8e-9923-745ea798cb8f-kube-api-access-brlrh\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528479 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-oauth-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.528506 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629700 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-trusted-ca-bundle\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-oauth-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629774 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brlrh\" (UniqueName: \"kubernetes.io/projected/b5d6ff30-da65-4a8e-9923-745ea798cb8f-kube-api-access-brlrh\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629803 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-oauth-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629821 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629843 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-service-ca\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.629877 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.631129 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.631187 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-oauth-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.631819 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-trusted-ca-bundle\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.632295 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5d6ff30-da65-4a8e-9923-745ea798cb8f-service-ca\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.633669 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-serving-cert\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.633691 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b5d6ff30-da65-4a8e-9923-745ea798cb8f-console-oauth-config\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.638437 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.645238 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brlrh\" (UniqueName: \"kubernetes.io/projected/b5d6ff30-da65-4a8e-9923-745ea798cb8f-kube-api-access-brlrh\") pod \"console-7bb44447b-b8xnt\" (UID: \"b5d6ff30-da65-4a8e-9923-745ea798cb8f\") " pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.687682 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" event={"ID":"a713c14e-bce4-44a3-940d-0a987cc6fa5b","Type":"ContainerStarted","Data":"554ff638b0ca662fa94ce2fea761bb54bff225199522ec5612edd645c8c70202"} Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.688638 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.688928 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sh4hc" event={"ID":"8109dd98-72c1-4542-a34c-1f5d911b5fd4","Type":"ContainerStarted","Data":"c51aca9e82f53c5d9b4a017dd3370e33563ee91bcb048e802736ad6f2fef1c56"} Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.689794 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" event={"ID":"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6","Type":"ContainerStarted","Data":"abc8fcb4fd52beeb0ed3464f3453f0f83d670ffc59c8ea9593e9b81d6483af63"} Sep 29 10:34:56 crc kubenswrapper[4727]: I0929 10:34:56.826226 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-shptq"] Sep 29 10:34:56 crc kubenswrapper[4727]: W0929 10:34:56.834041 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7655e999_03ae_45f0_a123_1efa69626f7b.slice/crio-d45688f12d360d2f648c0c085d910687059257baa704af140a35628ebfd9afad WatchSource:0}: Error finding container d45688f12d360d2f648c0c085d910687059257baa704af140a35628ebfd9afad: Status 404 returned error can't find the container with id d45688f12d360d2f648c0c085d910687059257baa704af140a35628ebfd9afad Sep 29 10:34:57 crc kubenswrapper[4727]: I0929 10:34:57.090987 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7bb44447b-b8xnt"] Sep 29 10:34:57 crc kubenswrapper[4727]: W0929 10:34:57.096143 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5d6ff30_da65_4a8e_9923_745ea798cb8f.slice/crio-90f4b3d87b9134646d92604d7b43ec3fab43e1691f148af0ef00339b8303aa9c WatchSource:0}: Error finding container 90f4b3d87b9134646d92604d7b43ec3fab43e1691f148af0ef00339b8303aa9c: Status 404 returned error can't find the container with id 90f4b3d87b9134646d92604d7b43ec3fab43e1691f148af0ef00339b8303aa9c Sep 29 10:34:57 crc kubenswrapper[4727]: I0929 10:34:57.695703 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" event={"ID":"7655e999-03ae-45f0-a123-1efa69626f7b","Type":"ContainerStarted","Data":"d45688f12d360d2f648c0c085d910687059257baa704af140a35628ebfd9afad"} Sep 29 10:34:57 crc kubenswrapper[4727]: I0929 10:34:57.697288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7bb44447b-b8xnt" event={"ID":"b5d6ff30-da65-4a8e-9923-745ea798cb8f","Type":"ContainerStarted","Data":"5f1d88fb643cb3a4eb7ee0fc9a7edc1cbeef49b8bb9fa809c78bab114da7edf7"} Sep 29 10:34:57 crc kubenswrapper[4727]: I0929 10:34:57.697517 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7bb44447b-b8xnt" event={"ID":"b5d6ff30-da65-4a8e-9923-745ea798cb8f","Type":"ContainerStarted","Data":"90f4b3d87b9134646d92604d7b43ec3fab43e1691f148af0ef00339b8303aa9c"} Sep 29 10:34:58 crc kubenswrapper[4727]: I0929 10:34:58.740601 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7bb44447b-b8xnt" podStartSLOduration=2.74058217 podStartE2EDuration="2.74058217s" podCreationTimestamp="2025-09-29 10:34:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:34:58.739292616 +0000 UTC m=+768.912605978" watchObservedRunningTime="2025-09-29 10:34:58.74058217 +0000 UTC m=+768.913895532" Sep 29 10:35:03 crc kubenswrapper[4727]: I0929 10:35:03.711658 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:35:03 crc kubenswrapper[4727]: I0929 10:35:03.712230 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" containerID="cri-o://7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3" gracePeriod=30 Sep 29 10:35:03 crc kubenswrapper[4727]: I0929 10:35:03.817974 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:35:03 crc kubenswrapper[4727]: I0929 10:35:03.818213 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" containerID="cri-o://ace6edc6233d2af7cd6f36b3fdcddc0fd67a146067a8963d3b0466a0e852d99a" gracePeriod=30 Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.525626 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.639729 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdxkx\" (UniqueName: \"kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx\") pod \"305b863d-a41d-437c-986b-04a2bb52fde9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.639808 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert\") pod \"305b863d-a41d-437c-986b-04a2bb52fde9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.639914 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca\") pod \"305b863d-a41d-437c-986b-04a2bb52fde9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.639952 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config\") pod \"305b863d-a41d-437c-986b-04a2bb52fde9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.639982 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles\") pod \"305b863d-a41d-437c-986b-04a2bb52fde9\" (UID: \"305b863d-a41d-437c-986b-04a2bb52fde9\") " Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.640659 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca" (OuterVolumeSpecName: "client-ca") pod "305b863d-a41d-437c-986b-04a2bb52fde9" (UID: "305b863d-a41d-437c-986b-04a2bb52fde9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.640729 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config" (OuterVolumeSpecName: "config") pod "305b863d-a41d-437c-986b-04a2bb52fde9" (UID: "305b863d-a41d-437c-986b-04a2bb52fde9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.640785 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "305b863d-a41d-437c-986b-04a2bb52fde9" (UID: "305b863d-a41d-437c-986b-04a2bb52fde9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.644424 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx" (OuterVolumeSpecName: "kube-api-access-wdxkx") pod "305b863d-a41d-437c-986b-04a2bb52fde9" (UID: "305b863d-a41d-437c-986b-04a2bb52fde9"). InnerVolumeSpecName "kube-api-access-wdxkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.645094 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "305b863d-a41d-437c-986b-04a2bb52fde9" (UID: "305b863d-a41d-437c-986b-04a2bb52fde9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.742924 4727 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.742965 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.742974 4727 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/305b863d-a41d-437c-986b-04a2bb52fde9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.742985 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdxkx\" (UniqueName: \"kubernetes.io/projected/305b863d-a41d-437c-986b-04a2bb52fde9-kube-api-access-wdxkx\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.742994 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/305b863d-a41d-437c-986b-04a2bb52fde9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.750496 4727 generic.go:334] "Generic (PLEG): container finished" podID="305b863d-a41d-437c-986b-04a2bb52fde9" containerID="7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3" exitCode=0 Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.750602 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.753405 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" event={"ID":"305b863d-a41d-437c-986b-04a2bb52fde9","Type":"ContainerDied","Data":"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3"} Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.753824 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r97m9" event={"ID":"305b863d-a41d-437c-986b-04a2bb52fde9","Type":"ContainerDied","Data":"6466e6b61047c567a2a27904b29c879371bc07f697fd1f1bea4ef0d034aa5c77"} Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.753877 4727 scope.go:117] "RemoveContainer" containerID="7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.758049 4727 generic.go:334] "Generic (PLEG): container finished" podID="357e2cd6-6aad-411f-a684-704708b9d291" containerID="ace6edc6233d2af7cd6f36b3fdcddc0fd67a146067a8963d3b0466a0e852d99a" exitCode=0 Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.758101 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" event={"ID":"357e2cd6-6aad-411f-a684-704708b9d291","Type":"ContainerDied","Data":"ace6edc6233d2af7cd6f36b3fdcddc0fd67a146067a8963d3b0466a0e852d99a"} Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.779634 4727 scope.go:117] "RemoveContainer" containerID="7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3" Sep 29 10:35:04 crc kubenswrapper[4727]: E0929 10:35:04.780021 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3\": container with ID starting with 7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3 not found: ID does not exist" containerID="7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.780060 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3"} err="failed to get container status \"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3\": rpc error: code = NotFound desc = could not find container \"7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3\": container with ID starting with 7e0e0e009fb65d2a6faaa48df202caef74f1c336605588b907107d00815821b3 not found: ID does not exist" Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.780269 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.787856 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r97m9"] Sep 29 10:35:04 crc kubenswrapper[4727]: I0929 10:35:04.921261 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.045928 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert\") pod \"357e2cd6-6aad-411f-a684-704708b9d291\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.045989 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca\") pod \"357e2cd6-6aad-411f-a684-704708b9d291\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.046047 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config\") pod \"357e2cd6-6aad-411f-a684-704708b9d291\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.046075 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj95w\" (UniqueName: \"kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w\") pod \"357e2cd6-6aad-411f-a684-704708b9d291\" (UID: \"357e2cd6-6aad-411f-a684-704708b9d291\") " Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.047753 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca" (OuterVolumeSpecName: "client-ca") pod "357e2cd6-6aad-411f-a684-704708b9d291" (UID: "357e2cd6-6aad-411f-a684-704708b9d291"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.048544 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config" (OuterVolumeSpecName: "config") pod "357e2cd6-6aad-411f-a684-704708b9d291" (UID: "357e2cd6-6aad-411f-a684-704708b9d291"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.053783 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w" (OuterVolumeSpecName: "kube-api-access-mj95w") pod "357e2cd6-6aad-411f-a684-704708b9d291" (UID: "357e2cd6-6aad-411f-a684-704708b9d291"). InnerVolumeSpecName "kube-api-access-mj95w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.053820 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "357e2cd6-6aad-411f-a684-704708b9d291" (UID: "357e2cd6-6aad-411f-a684-704708b9d291"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.116980 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" path="/var/lib/kubelet/pods/305b863d-a41d-437c-986b-04a2bb52fde9/volumes" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.147731 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.147765 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj95w\" (UniqueName: \"kubernetes.io/projected/357e2cd6-6aad-411f-a684-704708b9d291-kube-api-access-mj95w\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.147779 4727 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/357e2cd6-6aad-411f-a684-704708b9d291-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.147790 4727 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/357e2cd6-6aad-411f-a684-704708b9d291-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.484459 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc"] Sep 29 10:35:05 crc kubenswrapper[4727]: E0929 10:35:05.485515 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.485544 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: E0929 10:35:05.485580 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.485591 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.485751 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="305b863d-a41d-437c-986b-04a2bb52fde9" containerName="controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.485775 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="357e2cd6-6aad-411f-a684-704708b9d291" containerName="route-controller-manager" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.486628 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.489048 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-dd86c69bf-4cb6g"] Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.490208 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.496438 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.496729 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.496806 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.497107 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.497374 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.498277 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.505083 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.519479 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-dd86c69bf-4cb6g"] Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.526856 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc"] Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655543 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6225p\" (UniqueName: \"kubernetes.io/projected/0a366cab-6fac-4083-9f09-15a9b2454a84-kube-api-access-6225p\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655622 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-client-ca\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655666 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a366cab-6fac-4083-9f09-15a9b2454a84-serving-cert\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655705 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-config\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655750 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-config\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655771 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-serving-cert\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655792 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-client-ca\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.655890 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-proxy-ca-bundles\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.656083 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj9ff\" (UniqueName: \"kubernetes.io/projected/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-kube-api-access-mj9ff\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.758028 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6225p\" (UniqueName: \"kubernetes.io/projected/0a366cab-6fac-4083-9f09-15a9b2454a84-kube-api-access-6225p\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.758116 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-client-ca\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.758164 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a366cab-6fac-4083-9f09-15a9b2454a84-serving-cert\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759053 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-config\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759094 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-config\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759149 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-serving-cert\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759180 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-client-ca\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759238 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-proxy-ca-bundles\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759242 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-client-ca\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.759717 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj9ff\" (UniqueName: \"kubernetes.io/projected/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-kube-api-access-mj9ff\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.760318 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-config\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.760656 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a366cab-6fac-4083-9f09-15a9b2454a84-client-ca\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.760680 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-proxy-ca-bundles\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.761549 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-config\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.763822 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-serving-cert\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.767328 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" event={"ID":"7655e999-03ae-45f0-a123-1efa69626f7b","Type":"ContainerStarted","Data":"3ae85ffb1b8d228c0d0ffa1c349bdfb1c3705df3f28adeb49d62f6c624d915b7"} Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.768385 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.770445 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" event={"ID":"357e2cd6-6aad-411f-a684-704708b9d291","Type":"ContainerDied","Data":"a8d6c088981d6fb07f316741eabf436a9cab021621ef6812b0f5a8c5d20aaabc"} Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.770484 4727 scope.go:117] "RemoveContainer" containerID="ace6edc6233d2af7cd6f36b3fdcddc0fd67a146067a8963d3b0466a0e852d99a" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.770587 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.773924 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" event={"ID":"a713c14e-bce4-44a3-940d-0a987cc6fa5b","Type":"ContainerStarted","Data":"77ba82059e9b1d61208b50027f77b8b383e88901f7ec2e8f6a5193147e3f213b"} Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.775766 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a366cab-6fac-4083-9f09-15a9b2454a84-serving-cert\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.776151 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sh4hc" event={"ID":"8109dd98-72c1-4542-a34c-1f5d911b5fd4","Type":"ContainerStarted","Data":"528d1594b21da1d7b92db2596dad1350fccaecb1c54a2e2123f24500732e52ec"} Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.776783 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.784469 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6225p\" (UniqueName: \"kubernetes.io/projected/0a366cab-6fac-4083-9f09-15a9b2454a84-kube-api-access-6225p\") pod \"route-controller-manager-75f9d69c85-h49lc\" (UID: \"0a366cab-6fac-4083-9f09-15a9b2454a84\") " pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.784684 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" event={"ID":"8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6","Type":"ContainerStarted","Data":"cd8ef9bf711117f2b63a23e0febbf3ea295f212f86605f139d932f08df137473"} Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.786018 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj9ff\" (UniqueName: \"kubernetes.io/projected/d59e6838-4173-4c3c-a7d7-95dc95ff0bbf-kube-api-access-mj9ff\") pod \"controller-manager-dd86c69bf-4cb6g\" (UID: \"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf\") " pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.795694 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" podStartSLOduration=3.197564285 podStartE2EDuration="10.795678008s" podCreationTimestamp="2025-09-29 10:34:55 +0000 UTC" firstStartedPulling="2025-09-29 10:34:56.836391269 +0000 UTC m=+767.009704631" lastFinishedPulling="2025-09-29 10:35:04.434504992 +0000 UTC m=+774.607818354" observedRunningTime="2025-09-29 10:35:05.795588885 +0000 UTC m=+775.968902267" watchObservedRunningTime="2025-09-29 10:35:05.795678008 +0000 UTC m=+775.968991370" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.823433 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.823377 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-sh4hc" podStartSLOduration=2.149191755 podStartE2EDuration="10.82332124s" podCreationTimestamp="2025-09-29 10:34:55 +0000 UTC" firstStartedPulling="2025-09-29 10:34:56.117565136 +0000 UTC m=+766.290878488" lastFinishedPulling="2025-09-29 10:35:04.791694611 +0000 UTC m=+774.965007973" observedRunningTime="2025-09-29 10:35:05.815073207 +0000 UTC m=+775.988386569" watchObservedRunningTime="2025-09-29 10:35:05.82332124 +0000 UTC m=+775.996634602" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.833716 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-l7xjv" podStartSLOduration=2.818503872 podStartE2EDuration="10.833697686s" podCreationTimestamp="2025-09-29 10:34:55 +0000 UTC" firstStartedPulling="2025-09-29 10:34:56.419275437 +0000 UTC m=+766.592588799" lastFinishedPulling="2025-09-29 10:35:04.434469241 +0000 UTC m=+774.607782613" observedRunningTime="2025-09-29 10:35:05.83309312 +0000 UTC m=+776.006406502" watchObservedRunningTime="2025-09-29 10:35:05.833697686 +0000 UTC m=+776.007011048" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.833850 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.849070 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:35:05 crc kubenswrapper[4727]: I0929 10:35:05.858786 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fcr2d"] Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.110469 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-dd86c69bf-4cb6g"] Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.274889 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc"] Sep 29 10:35:06 crc kubenswrapper[4727]: W0929 10:35:06.288192 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a366cab_6fac_4083_9f09_15a9b2454a84.slice/crio-a65e3e5c14ffbd794e189774b7d3187f8375f217152952a9e1d4cfbecc9575ad WatchSource:0}: Error finding container a65e3e5c14ffbd794e189774b7d3187f8375f217152952a9e1d4cfbecc9575ad: Status 404 returned error can't find the container with id a65e3e5c14ffbd794e189774b7d3187f8375f217152952a9e1d4cfbecc9575ad Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.689049 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.689091 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.694402 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.796432 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" event={"ID":"0a366cab-6fac-4083-9f09-15a9b2454a84","Type":"ContainerStarted","Data":"4f02a73391db238ddc737673e53fc0236400aa3b6a708a395ddf16301b837fbc"} Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.796770 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" event={"ID":"0a366cab-6fac-4083-9f09-15a9b2454a84","Type":"ContainerStarted","Data":"a65e3e5c14ffbd794e189774b7d3187f8375f217152952a9e1d4cfbecc9575ad"} Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.796793 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.804609 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" event={"ID":"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf","Type":"ContainerStarted","Data":"bae4e179f8981ec3709f03a2077847e6b826a95314fc5302084418620a7ad68a"} Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.804651 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" event={"ID":"d59e6838-4173-4c3c-a7d7-95dc95ff0bbf","Type":"ContainerStarted","Data":"25df8f3cd4656f19b5f0d932f016e5d77a70e35457ab3678884a54d15d7fa669"} Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.808130 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7bb44447b-b8xnt" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.863317 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" podStartSLOduration=3.863295292 podStartE2EDuration="3.863295292s" podCreationTimestamp="2025-09-29 10:35:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:35:06.823031045 +0000 UTC m=+776.996344407" watchObservedRunningTime="2025-09-29 10:35:06.863295292 +0000 UTC m=+777.036608654" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.864157 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" podStartSLOduration=3.8641504639999997 podStartE2EDuration="3.864150464s" podCreationTimestamp="2025-09-29 10:35:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:35:06.862275926 +0000 UTC m=+777.035589308" watchObservedRunningTime="2025-09-29 10:35:06.864150464 +0000 UTC m=+777.037463826" Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.945202 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:35:06 crc kubenswrapper[4727]: I0929 10:35:06.967800 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75f9d69c85-h49lc" Sep 29 10:35:07 crc kubenswrapper[4727]: I0929 10:35:07.120436 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="357e2cd6-6aad-411f-a684-704708b9d291" path="/var/lib/kubelet/pods/357e2cd6-6aad-411f-a684-704708b9d291/volumes" Sep 29 10:35:07 crc kubenswrapper[4727]: I0929 10:35:07.809638 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:07 crc kubenswrapper[4727]: I0929 10:35:07.813788 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-dd86c69bf-4cb6g" Sep 29 10:35:11 crc kubenswrapper[4727]: I0929 10:35:11.115349 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-sh4hc" Sep 29 10:35:11 crc kubenswrapper[4727]: I0929 10:35:11.794408 4727 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 10:35:11 crc kubenswrapper[4727]: I0929 10:35:11.834267 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" event={"ID":"a713c14e-bce4-44a3-940d-0a987cc6fa5b","Type":"ContainerStarted","Data":"0138e936889d6e29387d9c5a23203bd0f4d304054228ddd895722bc2a7ef9a6e"} Sep 29 10:35:11 crc kubenswrapper[4727]: I0929 10:35:11.859598 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2hx48" podStartSLOduration=2.407914998 podStartE2EDuration="16.859581037s" podCreationTimestamp="2025-09-29 10:34:55 +0000 UTC" firstStartedPulling="2025-09-29 10:34:56.240228425 +0000 UTC m=+766.413541787" lastFinishedPulling="2025-09-29 10:35:10.691894464 +0000 UTC m=+780.865207826" observedRunningTime="2025-09-29 10:35:11.858690074 +0000 UTC m=+782.032003446" watchObservedRunningTime="2025-09-29 10:35:11.859581037 +0000 UTC m=+782.032894399" Sep 29 10:35:16 crc kubenswrapper[4727]: I0929 10:35:16.643652 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-shptq" Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.246694 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.246761 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.246816 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.249141 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.249223 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090" gracePeriod=600 Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.886152 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090" exitCode=0 Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.886246 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090"} Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.886730 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435"} Sep 29 10:35:19 crc kubenswrapper[4727]: I0929 10:35:19.886761 4727 scope.go:117] "RemoveContainer" containerID="6dbdd1cc7953aad539dbc049d2a3bfb719b1b72f16fec786df578a8c9e455a16" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.059105 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst"] Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.061071 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.064951 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.076673 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst"] Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.123739 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctkw\" (UniqueName: \"kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.123805 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.123848 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.225741 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctkw\" (UniqueName: \"kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.225809 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.225854 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.226434 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.226719 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.278318 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctkw\" (UniqueName: \"kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.379623 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.799235 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst"] Sep 29 10:35:30 crc kubenswrapper[4727]: I0929 10:35:30.969525 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" event={"ID":"5d62033f-fb65-4d2f-b853-d241790d6bca","Type":"ContainerStarted","Data":"5ada1764a51ad61f3095182ca892c3abce4eb6344b22e330217d569987df4869"} Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.757850 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.759280 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.770421 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.859832 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.859923 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbz9b\" (UniqueName: \"kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.859959 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.961740 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.961826 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbz9b\" (UniqueName: \"kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.961863 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.962500 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.962533 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.978605 4727 generic.go:334] "Generic (PLEG): container finished" podID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerID="b0589fcb209f051abf32f5b4a975803f836ff44637479b2cae3c9ea2958c1bee" exitCode=0 Sep 29 10:35:31 crc kubenswrapper[4727]: I0929 10:35:31.978671 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" event={"ID":"5d62033f-fb65-4d2f-b853-d241790d6bca","Type":"ContainerDied","Data":"b0589fcb209f051abf32f5b4a975803f836ff44637479b2cae3c9ea2958c1bee"} Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.007913 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbz9b\" (UniqueName: \"kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b\") pod \"redhat-operators-d9ztl\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.023167 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-fkfsn" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" containerID="cri-o://926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab" gracePeriod=15 Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.115145 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.543573 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fkfsn_7451dc02-38ba-47f9-8bca-5ec1ca2ffef5/console/0.log" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.543934 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.569193 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.569232 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.569261 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.569306 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570320 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570333 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570430 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtpm4\" (UniqueName: \"kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570524 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config" (OuterVolumeSpecName: "console-config") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570794 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.570868 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config\") pod \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\" (UID: \"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5\") " Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.571139 4727 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.571157 4727 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.571167 4727 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.571774 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca" (OuterVolumeSpecName: "service-ca") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.576087 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4" (OuterVolumeSpecName: "kube-api-access-mtpm4") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "kube-api-access-mtpm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.576245 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.584244 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" (UID: "7451dc02-38ba-47f9-8bca-5ec1ca2ffef5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.659080 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:32 crc kubenswrapper[4727]: W0929 10:35:32.665670 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod837eadd9_1220_45a6_b658_ecc96be83779.slice/crio-9d3ed3e2f6492de550111c5cff054bba51aa00cb816dc95e532438ce82336166 WatchSource:0}: Error finding container 9d3ed3e2f6492de550111c5cff054bba51aa00cb816dc95e532438ce82336166: Status 404 returned error can't find the container with id 9d3ed3e2f6492de550111c5cff054bba51aa00cb816dc95e532438ce82336166 Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.672562 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtpm4\" (UniqueName: \"kubernetes.io/projected/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-kube-api-access-mtpm4\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.672596 4727 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.672610 4727 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.672627 4727 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987701 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fkfsn_7451dc02-38ba-47f9-8bca-5ec1ca2ffef5/console/0.log" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987759 4727 generic.go:334] "Generic (PLEG): container finished" podID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerID="926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab" exitCode=2 Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987830 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fkfsn" event={"ID":"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5","Type":"ContainerDied","Data":"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab"} Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987875 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fkfsn" event={"ID":"7451dc02-38ba-47f9-8bca-5ec1ca2ffef5","Type":"ContainerDied","Data":"d9a0154e7f8d1c04ed7fceef7a1218fb4fec95c7923e37f198ced4b7291b061b"} Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987898 4727 scope.go:117] "RemoveContainer" containerID="926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.987956 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fkfsn" Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.989685 4727 generic.go:334] "Generic (PLEG): container finished" podID="837eadd9-1220-45a6-b658-ecc96be83779" containerID="a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63" exitCode=0 Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.989721 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerDied","Data":"a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63"} Sep 29 10:35:32 crc kubenswrapper[4727]: I0929 10:35:32.989748 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerStarted","Data":"9d3ed3e2f6492de550111c5cff054bba51aa00cb816dc95e532438ce82336166"} Sep 29 10:35:33 crc kubenswrapper[4727]: I0929 10:35:33.008603 4727 scope.go:117] "RemoveContainer" containerID="926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab" Sep 29 10:35:33 crc kubenswrapper[4727]: E0929 10:35:33.010890 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab\": container with ID starting with 926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab not found: ID does not exist" containerID="926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab" Sep 29 10:35:33 crc kubenswrapper[4727]: I0929 10:35:33.010923 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab"} err="failed to get container status \"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab\": rpc error: code = NotFound desc = could not find container \"926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab\": container with ID starting with 926c66fee908f83fb8f9dacebae03aab7e5b7da33eae4aed3134b599c7447cab not found: ID does not exist" Sep 29 10:35:33 crc kubenswrapper[4727]: I0929 10:35:33.057313 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:35:33 crc kubenswrapper[4727]: I0929 10:35:33.080933 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-fkfsn"] Sep 29 10:35:33 crc kubenswrapper[4727]: I0929 10:35:33.116537 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" path="/var/lib/kubelet/pods/7451dc02-38ba-47f9-8bca-5ec1ca2ffef5/volumes" Sep 29 10:35:35 crc kubenswrapper[4727]: I0929 10:35:35.020161 4727 generic.go:334] "Generic (PLEG): container finished" podID="837eadd9-1220-45a6-b658-ecc96be83779" containerID="11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852" exitCode=0 Sep 29 10:35:35 crc kubenswrapper[4727]: I0929 10:35:35.020259 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerDied","Data":"11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852"} Sep 29 10:35:35 crc kubenswrapper[4727]: I0929 10:35:35.024431 4727 generic.go:334] "Generic (PLEG): container finished" podID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerID="144771afcf43e1b16c0580a9005e5fb8b8238e65f91e1ec78645264b4064d7d7" exitCode=0 Sep 29 10:35:35 crc kubenswrapper[4727]: I0929 10:35:35.024502 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" event={"ID":"5d62033f-fb65-4d2f-b853-d241790d6bca","Type":"ContainerDied","Data":"144771afcf43e1b16c0580a9005e5fb8b8238e65f91e1ec78645264b4064d7d7"} Sep 29 10:35:36 crc kubenswrapper[4727]: I0929 10:35:36.032110 4727 generic.go:334] "Generic (PLEG): container finished" podID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerID="a451aa18cc4f5532b2b7180c625f23d0e782f55093436127840aa1d47dc74e79" exitCode=0 Sep 29 10:35:36 crc kubenswrapper[4727]: I0929 10:35:36.032189 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" event={"ID":"5d62033f-fb65-4d2f-b853-d241790d6bca","Type":"ContainerDied","Data":"a451aa18cc4f5532b2b7180c625f23d0e782f55093436127840aa1d47dc74e79"} Sep 29 10:35:36 crc kubenswrapper[4727]: I0929 10:35:36.035900 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerStarted","Data":"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1"} Sep 29 10:35:36 crc kubenswrapper[4727]: I0929 10:35:36.069610 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d9ztl" podStartSLOduration=2.641134918 podStartE2EDuration="5.06957746s" podCreationTimestamp="2025-09-29 10:35:31 +0000 UTC" firstStartedPulling="2025-09-29 10:35:32.991255041 +0000 UTC m=+803.164568403" lastFinishedPulling="2025-09-29 10:35:35.419697583 +0000 UTC m=+805.593010945" observedRunningTime="2025-09-29 10:35:36.068691587 +0000 UTC m=+806.242004969" watchObservedRunningTime="2025-09-29 10:35:36.06957746 +0000 UTC m=+806.242890822" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.394398 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.448954 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rctkw\" (UniqueName: \"kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw\") pod \"5d62033f-fb65-4d2f-b853-d241790d6bca\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.449008 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util\") pod \"5d62033f-fb65-4d2f-b853-d241790d6bca\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.449056 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle\") pod \"5d62033f-fb65-4d2f-b853-d241790d6bca\" (UID: \"5d62033f-fb65-4d2f-b853-d241790d6bca\") " Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.450156 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle" (OuterVolumeSpecName: "bundle") pod "5d62033f-fb65-4d2f-b853-d241790d6bca" (UID: "5d62033f-fb65-4d2f-b853-d241790d6bca"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.450301 4727 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.456686 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw" (OuterVolumeSpecName: "kube-api-access-rctkw") pod "5d62033f-fb65-4d2f-b853-d241790d6bca" (UID: "5d62033f-fb65-4d2f-b853-d241790d6bca"). InnerVolumeSpecName "kube-api-access-rctkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.461501 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util" (OuterVolumeSpecName: "util") pod "5d62033f-fb65-4d2f-b853-d241790d6bca" (UID: "5d62033f-fb65-4d2f-b853-d241790d6bca"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.550864 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rctkw\" (UniqueName: \"kubernetes.io/projected/5d62033f-fb65-4d2f-b853-d241790d6bca-kube-api-access-rctkw\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:37 crc kubenswrapper[4727]: I0929 10:35:37.550913 4727 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d62033f-fb65-4d2f-b853-d241790d6bca-util\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:38 crc kubenswrapper[4727]: I0929 10:35:38.050082 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" event={"ID":"5d62033f-fb65-4d2f-b853-d241790d6bca","Type":"ContainerDied","Data":"5ada1764a51ad61f3095182ca892c3abce4eb6344b22e330217d569987df4869"} Sep 29 10:35:38 crc kubenswrapper[4727]: I0929 10:35:38.050125 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ada1764a51ad61f3095182ca892c3abce4eb6344b22e330217d569987df4869" Sep 29 10:35:38 crc kubenswrapper[4727]: I0929 10:35:38.050217 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst" Sep 29 10:35:42 crc kubenswrapper[4727]: I0929 10:35:42.116073 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:42 crc kubenswrapper[4727]: I0929 10:35:42.117309 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:42 crc kubenswrapper[4727]: I0929 10:35:42.175508 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:43 crc kubenswrapper[4727]: I0929 10:35:43.121604 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:44 crc kubenswrapper[4727]: I0929 10:35:44.143816 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.096297 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d9ztl" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="registry-server" containerID="cri-o://95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1" gracePeriod=2 Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.512685 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.650499 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities\") pod \"837eadd9-1220-45a6-b658-ecc96be83779\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.650575 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbz9b\" (UniqueName: \"kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b\") pod \"837eadd9-1220-45a6-b658-ecc96be83779\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.650673 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content\") pod \"837eadd9-1220-45a6-b658-ecc96be83779\" (UID: \"837eadd9-1220-45a6-b658-ecc96be83779\") " Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.651598 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities" (OuterVolumeSpecName: "utilities") pod "837eadd9-1220-45a6-b658-ecc96be83779" (UID: "837eadd9-1220-45a6-b658-ecc96be83779"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.656068 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b" (OuterVolumeSpecName: "kube-api-access-gbz9b") pod "837eadd9-1220-45a6-b658-ecc96be83779" (UID: "837eadd9-1220-45a6-b658-ecc96be83779"). InnerVolumeSpecName "kube-api-access-gbz9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.752774 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:45 crc kubenswrapper[4727]: I0929 10:35:45.753058 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbz9b\" (UniqueName: \"kubernetes.io/projected/837eadd9-1220-45a6-b658-ecc96be83779-kube-api-access-gbz9b\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.106816 4727 generic.go:334] "Generic (PLEG): container finished" podID="837eadd9-1220-45a6-b658-ecc96be83779" containerID="95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1" exitCode=0 Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.106861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerDied","Data":"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1"} Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.106892 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9ztl" event={"ID":"837eadd9-1220-45a6-b658-ecc96be83779","Type":"ContainerDied","Data":"9d3ed3e2f6492de550111c5cff054bba51aa00cb816dc95e532438ce82336166"} Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.106911 4727 scope.go:117] "RemoveContainer" containerID="95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.107038 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9ztl" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.122884 4727 scope.go:117] "RemoveContainer" containerID="11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.141292 4727 scope.go:117] "RemoveContainer" containerID="a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.161236 4727 scope.go:117] "RemoveContainer" containerID="95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1" Sep 29 10:35:46 crc kubenswrapper[4727]: E0929 10:35:46.161964 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1\": container with ID starting with 95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1 not found: ID does not exist" containerID="95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.162029 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1"} err="failed to get container status \"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1\": rpc error: code = NotFound desc = could not find container \"95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1\": container with ID starting with 95f4bded46986c4a70d9b18cbc9cdd668ba9b3af436444d1131cc05843830cf1 not found: ID does not exist" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.162052 4727 scope.go:117] "RemoveContainer" containerID="11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852" Sep 29 10:35:46 crc kubenswrapper[4727]: E0929 10:35:46.162284 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852\": container with ID starting with 11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852 not found: ID does not exist" containerID="11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.162313 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852"} err="failed to get container status \"11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852\": rpc error: code = NotFound desc = could not find container \"11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852\": container with ID starting with 11fd6cce555ffb6502fdc6782f43a60439271e21981bc10ff85fbadde9bb6852 not found: ID does not exist" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.162330 4727 scope.go:117] "RemoveContainer" containerID="a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63" Sep 29 10:35:46 crc kubenswrapper[4727]: E0929 10:35:46.162595 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63\": container with ID starting with a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63 not found: ID does not exist" containerID="a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63" Sep 29 10:35:46 crc kubenswrapper[4727]: I0929 10:35:46.162623 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63"} err="failed to get container status \"a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63\": rpc error: code = NotFound desc = could not find container \"a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63\": container with ID starting with a22f6f67274a28f2a3feb49be5b67424d40bbb3b230e8d82034a5d306c702d63 not found: ID does not exist" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.560983 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz"] Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561719 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="registry-server" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561737 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="registry-server" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561748 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561757 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561791 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="pull" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561802 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="pull" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561823 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="extract-content" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561832 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="extract-content" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561845 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="util" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561877 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="util" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561897 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="extract-utilities" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561906 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="extract-utilities" Sep 29 10:35:47 crc kubenswrapper[4727]: E0929 10:35:47.561916 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="extract" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.561924 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="extract" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.562126 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d62033f-fb65-4d2f-b853-d241790d6bca" containerName="extract" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.562150 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="7451dc02-38ba-47f9-8bca-5ec1ca2ffef5" containerName="console" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.562162 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="837eadd9-1220-45a6-b658-ecc96be83779" containerName="registry-server" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.562910 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.566653 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.566656 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.566796 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.567586 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.581371 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz"] Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.585514 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-nrknf" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.679124 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-webhook-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.679179 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-apiservice-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.679354 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqfrv\" (UniqueName: \"kubernetes.io/projected/9e45a861-9d6a-43af-9235-33e4cea43955-kube-api-access-fqfrv\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.780737 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-webhook-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.780785 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-apiservice-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.780824 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqfrv\" (UniqueName: \"kubernetes.io/projected/9e45a861-9d6a-43af-9235-33e4cea43955-kube-api-access-fqfrv\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.789425 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-webhook-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.800124 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqfrv\" (UniqueName: \"kubernetes.io/projected/9e45a861-9d6a-43af-9235-33e4cea43955-kube-api-access-fqfrv\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.806080 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9e45a861-9d6a-43af-9235-33e4cea43955-apiservice-cert\") pod \"metallb-operator-controller-manager-86747676bf-9fmjz\" (UID: \"9e45a861-9d6a-43af-9235-33e4cea43955\") " pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.878385 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.980495 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8"] Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.981558 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.989676 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.989744 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 29 10:35:47 crc kubenswrapper[4727]: I0929 10:35:47.989845 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-g4rwp" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.025770 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8"] Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.054201 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "837eadd9-1220-45a6-b658-ecc96be83779" (UID: "837eadd9-1220-45a6-b658-ecc96be83779"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.084132 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-webhook-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.084198 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crgjn\" (UniqueName: \"kubernetes.io/projected/5c0a38f9-cb25-4466-9d31-4d6bf3943989-kube-api-access-crgjn\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.084249 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.084298 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/837eadd9-1220-45a6-b658-ecc96be83779-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.185303 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.185395 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-webhook-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.185434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crgjn\" (UniqueName: \"kubernetes.io/projected/5c0a38f9-cb25-4466-9d31-4d6bf3943989-kube-api-access-crgjn\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.192062 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-apiservice-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.207979 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c0a38f9-cb25-4466-9d31-4d6bf3943989-webhook-cert\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.221521 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crgjn\" (UniqueName: \"kubernetes.io/projected/5c0a38f9-cb25-4466-9d31-4d6bf3943989-kube-api-access-crgjn\") pod \"metallb-operator-webhook-server-5b8c57d74f-nnlj8\" (UID: \"5c0a38f9-cb25-4466-9d31-4d6bf3943989\") " pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.235203 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.238585 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d9ztl"] Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.316912 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.350436 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz"] Sep 29 10:35:48 crc kubenswrapper[4727]: I0929 10:35:48.741534 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8"] Sep 29 10:35:48 crc kubenswrapper[4727]: W0929 10:35:48.746680 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c0a38f9_cb25_4466_9d31_4d6bf3943989.slice/crio-cb539fb274ad886a0b33f01781f783fb9ad1119ad527126729cd9666dcb39f07 WatchSource:0}: Error finding container cb539fb274ad886a0b33f01781f783fb9ad1119ad527126729cd9666dcb39f07: Status 404 returned error can't find the container with id cb539fb274ad886a0b33f01781f783fb9ad1119ad527126729cd9666dcb39f07 Sep 29 10:35:49 crc kubenswrapper[4727]: I0929 10:35:49.115509 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="837eadd9-1220-45a6-b658-ecc96be83779" path="/var/lib/kubelet/pods/837eadd9-1220-45a6-b658-ecc96be83779/volumes" Sep 29 10:35:49 crc kubenswrapper[4727]: I0929 10:35:49.130937 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" event={"ID":"5c0a38f9-cb25-4466-9d31-4d6bf3943989","Type":"ContainerStarted","Data":"cb539fb274ad886a0b33f01781f783fb9ad1119ad527126729cd9666dcb39f07"} Sep 29 10:35:49 crc kubenswrapper[4727]: I0929 10:35:49.132122 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" event={"ID":"9e45a861-9d6a-43af-9235-33e4cea43955","Type":"ContainerStarted","Data":"7dfd8fa06fa17d4ad0735d9b090dfdb83e73184721679cd6802d927c977b3da1"} Sep 29 10:35:54 crc kubenswrapper[4727]: I0929 10:35:54.167514 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" event={"ID":"9e45a861-9d6a-43af-9235-33e4cea43955","Type":"ContainerStarted","Data":"b8e79ada25897f78f35b137d27756ef95050446c5ddbf88fcd2fbc93e9602e37"} Sep 29 10:35:54 crc kubenswrapper[4727]: I0929 10:35:54.168116 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:35:54 crc kubenswrapper[4727]: I0929 10:35:54.187126 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" podStartSLOduration=1.8636105189999999 podStartE2EDuration="7.187103929s" podCreationTimestamp="2025-09-29 10:35:47 +0000 UTC" firstStartedPulling="2025-09-29 10:35:48.360499401 +0000 UTC m=+818.533812763" lastFinishedPulling="2025-09-29 10:35:53.683992811 +0000 UTC m=+823.857306173" observedRunningTime="2025-09-29 10:35:54.185168899 +0000 UTC m=+824.358482271" watchObservedRunningTime="2025-09-29 10:35:54.187103929 +0000 UTC m=+824.360417291" Sep 29 10:35:56 crc kubenswrapper[4727]: I0929 10:35:56.181200 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" event={"ID":"5c0a38f9-cb25-4466-9d31-4d6bf3943989","Type":"ContainerStarted","Data":"989310c6d0841bad0ab47b5f4ad536f393d18881e980c50037ae77004e9e6e0e"} Sep 29 10:35:56 crc kubenswrapper[4727]: I0929 10:35:56.181642 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:35:56 crc kubenswrapper[4727]: I0929 10:35:56.209768 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" podStartSLOduration=2.790072607 podStartE2EDuration="9.209747299s" podCreationTimestamp="2025-09-29 10:35:47 +0000 UTC" firstStartedPulling="2025-09-29 10:35:48.752144277 +0000 UTC m=+818.925457639" lastFinishedPulling="2025-09-29 10:35:55.171818969 +0000 UTC m=+825.345132331" observedRunningTime="2025-09-29 10:35:56.207657306 +0000 UTC m=+826.380970668" watchObservedRunningTime="2025-09-29 10:35:56.209747299 +0000 UTC m=+826.383060661" Sep 29 10:36:08 crc kubenswrapper[4727]: I0929 10:36:08.321700 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b8c57d74f-nnlj8" Sep 29 10:36:27 crc kubenswrapper[4727]: I0929 10:36:27.881181 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-86747676bf-9fmjz" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.715526 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jrlvl"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.720084 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.720947 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.720939 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.725182 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.725791 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.726397 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.726453 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xp9s9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.758142 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.820636 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-vh9j9"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.821766 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.823980 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.824772 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-mrwqk" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.824934 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.828695 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.843458 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-b8fqx"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.844990 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.848580 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852091 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metallb-excludel2\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852149 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-metrics-certs\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852186 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-957hb\" (UniqueName: \"kubernetes.io/projected/424bbcbe-8ea1-491c-b1b0-55285bff680e-kube-api-access-957hb\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852225 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics-certs\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852259 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-reloader\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852291 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-sockets\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852349 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-conf\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852378 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpfdk\" (UniqueName: \"kubernetes.io/projected/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-kube-api-access-vpfdk\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852416 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852454 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852487 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-cert\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852524 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wwmb\" (UniqueName: \"kubernetes.io/projected/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-kube-api-access-6wwmb\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852551 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-startup\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852579 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/424bbcbe-8ea1-491c-b1b0-55285bff680e-cert\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852609 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.852640 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hcfp\" (UniqueName: \"kubernetes.io/projected/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-kube-api-access-7hcfp\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.873025 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-b8fqx"] Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953421 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953470 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953495 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-cert\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953516 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-startup\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953531 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wwmb\" (UniqueName: \"kubernetes.io/projected/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-kube-api-access-6wwmb\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953548 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/424bbcbe-8ea1-491c-b1b0-55285bff680e-cert\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953566 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953582 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hcfp\" (UniqueName: \"kubernetes.io/projected/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-kube-api-access-7hcfp\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953598 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metallb-excludel2\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953619 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-metrics-certs\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953634 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-957hb\" (UniqueName: \"kubernetes.io/projected/424bbcbe-8ea1-491c-b1b0-55285bff680e-kube-api-access-957hb\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953657 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics-certs\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953674 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-reloader\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953691 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-sockets\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953722 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-conf\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.953739 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpfdk\" (UniqueName: \"kubernetes.io/projected/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-kube-api-access-vpfdk\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: E0929 10:36:28.953788 4727 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 10:36:28 crc kubenswrapper[4727]: E0929 10:36:28.953935 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist podName:fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e nodeName:}" failed. No retries permitted until 2025-09-29 10:36:29.453895082 +0000 UTC m=+859.627208444 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist") pod "speaker-vh9j9" (UID: "fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e") : secret "metallb-memberlist" not found Sep 29 10:36:28 crc kubenswrapper[4727]: E0929 10:36:28.954523 4727 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Sep 29 10:36:28 crc kubenswrapper[4727]: E0929 10:36:28.954579 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs podName:fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e nodeName:}" failed. No retries permitted until 2025-09-29 10:36:29.454562489 +0000 UTC m=+859.627875851 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs") pod "speaker-vh9j9" (UID: "fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e") : secret "speaker-certs-secret" not found Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.955467 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metallb-excludel2\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.955847 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-reloader\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.955994 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.956197 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-startup\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.958491 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-conf\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.959255 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-frr-sockets\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.959273 4727 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.963313 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/424bbcbe-8ea1-491c-b1b0-55285bff680e-cert\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.965899 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-metrics-certs\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.967075 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-cert\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.970046 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-metrics-certs\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.973399 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hcfp\" (UniqueName: \"kubernetes.io/projected/5a5b2bf7-f24e-494d-a0fc-11c6f90737bd-kube-api-access-7hcfp\") pod \"controller-5d688f5ffc-b8fqx\" (UID: \"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd\") " pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.974517 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wwmb\" (UniqueName: \"kubernetes.io/projected/fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2-kube-api-access-6wwmb\") pod \"frr-k8s-jrlvl\" (UID: \"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2\") " pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.994299 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-957hb\" (UniqueName: \"kubernetes.io/projected/424bbcbe-8ea1-491c-b1b0-55285bff680e-kube-api-access-957hb\") pod \"frr-k8s-webhook-server-5478bdb765-6pmt2\" (UID: \"424bbcbe-8ea1-491c-b1b0-55285bff680e\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:28 crc kubenswrapper[4727]: I0929 10:36:28.999233 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpfdk\" (UniqueName: \"kubernetes.io/projected/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-kube-api-access-vpfdk\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.061004 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.068018 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.163075 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.464925 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:29 crc kubenswrapper[4727]: E0929 10:36:29.465375 4727 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 10:36:29 crc kubenswrapper[4727]: E0929 10:36:29.465736 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist podName:fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e nodeName:}" failed. No retries permitted until 2025-09-29 10:36:30.465712873 +0000 UTC m=+860.639026225 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist") pod "speaker-vh9j9" (UID: "fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e") : secret "metallb-memberlist" not found Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.465391 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.472309 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-metrics-certs\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.486013 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2"] Sep 29 10:36:29 crc kubenswrapper[4727]: W0929 10:36:29.494944 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod424bbcbe_8ea1_491c_b1b0_55285bff680e.slice/crio-c20db99da3d3ffefa334716199fda8f5808b60cb08592b0edb5e006b1dc2cf36 WatchSource:0}: Error finding container c20db99da3d3ffefa334716199fda8f5808b60cb08592b0edb5e006b1dc2cf36: Status 404 returned error can't find the container with id c20db99da3d3ffefa334716199fda8f5808b60cb08592b0edb5e006b1dc2cf36 Sep 29 10:36:29 crc kubenswrapper[4727]: I0929 10:36:29.716382 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-b8fqx"] Sep 29 10:36:29 crc kubenswrapper[4727]: W0929 10:36:29.726703 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a5b2bf7_f24e_494d_a0fc_11c6f90737bd.slice/crio-4bccfea6d61cf5e471779f04000fc39bd6bc3e6e46fa0a1d34da366b4dc2f263 WatchSource:0}: Error finding container 4bccfea6d61cf5e471779f04000fc39bd6bc3e6e46fa0a1d34da366b4dc2f263: Status 404 returned error can't find the container with id 4bccfea6d61cf5e471779f04000fc39bd6bc3e6e46fa0a1d34da366b4dc2f263 Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.401918 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-b8fqx" event={"ID":"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd","Type":"ContainerStarted","Data":"3c1b203b62f68cba7fdd427cd160093b660f695434f5835e4d0b34b125849556"} Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.402288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-b8fqx" event={"ID":"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd","Type":"ContainerStarted","Data":"4bccfea6d61cf5e471779f04000fc39bd6bc3e6e46fa0a1d34da366b4dc2f263"} Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.403308 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" event={"ID":"424bbcbe-8ea1-491c-b1b0-55285bff680e","Type":"ContainerStarted","Data":"c20db99da3d3ffefa334716199fda8f5808b60cb08592b0edb5e006b1dc2cf36"} Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.404622 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"4c4379566db5e74becdb6ea888a7d86122c4870965df93d3292e192764071439"} Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.480968 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.487886 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e-memberlist\") pod \"speaker-vh9j9\" (UID: \"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e\") " pod="metallb-system/speaker-vh9j9" Sep 29 10:36:30 crc kubenswrapper[4727]: I0929 10:36:30.636945 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vh9j9" Sep 29 10:36:30 crc kubenswrapper[4727]: W0929 10:36:30.664919 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa619b15_1b04_4ad9_bf3d_abfd4aee5b2e.slice/crio-0e44c26cdbd4dfe9483f5507ebcf348d67809ec1319f4debb81382ad84f53792 WatchSource:0}: Error finding container 0e44c26cdbd4dfe9483f5507ebcf348d67809ec1319f4debb81382ad84f53792: Status 404 returned error can't find the container with id 0e44c26cdbd4dfe9483f5507ebcf348d67809ec1319f4debb81382ad84f53792 Sep 29 10:36:31 crc kubenswrapper[4727]: I0929 10:36:31.412809 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vh9j9" event={"ID":"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e","Type":"ContainerStarted","Data":"a7fed2a06decfbd3e45031847b07ef172eafac6653e0af8488fa43b67dfa614c"} Sep 29 10:36:31 crc kubenswrapper[4727]: I0929 10:36:31.413136 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vh9j9" event={"ID":"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e","Type":"ContainerStarted","Data":"0e44c26cdbd4dfe9483f5507ebcf348d67809ec1319f4debb81382ad84f53792"} Sep 29 10:36:31 crc kubenswrapper[4727]: I0929 10:36:31.415693 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-b8fqx" event={"ID":"5a5b2bf7-f24e-494d-a0fc-11c6f90737bd","Type":"ContainerStarted","Data":"f0ddfb97ebb95a7a34b84d08a09667999d589d29133e061d3303546c8ce027c2"} Sep 29 10:36:31 crc kubenswrapper[4727]: I0929 10:36:31.415818 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:31 crc kubenswrapper[4727]: I0929 10:36:31.453956 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-b8fqx" podStartSLOduration=3.453927638 podStartE2EDuration="3.453927638s" podCreationTimestamp="2025-09-29 10:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:36:31.450272754 +0000 UTC m=+861.623586146" watchObservedRunningTime="2025-09-29 10:36:31.453927638 +0000 UTC m=+861.627241010" Sep 29 10:36:32 crc kubenswrapper[4727]: I0929 10:36:32.424646 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vh9j9" event={"ID":"fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e","Type":"ContainerStarted","Data":"d6b406a2bb349bdbf4ebed47a08b70837cb11d574f239ec1c1ab56fb93a72503"} Sep 29 10:36:32 crc kubenswrapper[4727]: I0929 10:36:32.424864 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-vh9j9" Sep 29 10:36:32 crc kubenswrapper[4727]: I0929 10:36:32.449749 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-vh9j9" podStartSLOduration=4.449720333 podStartE2EDuration="4.449720333s" podCreationTimestamp="2025-09-29 10:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:36:32.445991677 +0000 UTC m=+862.619305039" watchObservedRunningTime="2025-09-29 10:36:32.449720333 +0000 UTC m=+862.623033695" Sep 29 10:36:38 crc kubenswrapper[4727]: I0929 10:36:38.465045 4727 generic.go:334] "Generic (PLEG): container finished" podID="fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2" containerID="9fe93d307e7ae070376942048b6135994b5575c3bc618849f5e3404c27842934" exitCode=0 Sep 29 10:36:38 crc kubenswrapper[4727]: I0929 10:36:38.465153 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerDied","Data":"9fe93d307e7ae070376942048b6135994b5575c3bc618849f5e3404c27842934"} Sep 29 10:36:38 crc kubenswrapper[4727]: I0929 10:36:38.469680 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" event={"ID":"424bbcbe-8ea1-491c-b1b0-55285bff680e","Type":"ContainerStarted","Data":"8bec06af3b6004d1e4b14771c89044e59e9d2a3823bcf8b03a476dbc6a692f34"} Sep 29 10:36:38 crc kubenswrapper[4727]: I0929 10:36:38.469858 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:38 crc kubenswrapper[4727]: I0929 10:36:38.518355 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" podStartSLOduration=2.044125089 podStartE2EDuration="10.518318674s" podCreationTimestamp="2025-09-29 10:36:28 +0000 UTC" firstStartedPulling="2025-09-29 10:36:29.498255971 +0000 UTC m=+859.671569333" lastFinishedPulling="2025-09-29 10:36:37.972449566 +0000 UTC m=+868.145762918" observedRunningTime="2025-09-29 10:36:38.512154295 +0000 UTC m=+868.685467657" watchObservedRunningTime="2025-09-29 10:36:38.518318674 +0000 UTC m=+868.691632036" Sep 29 10:36:39 crc kubenswrapper[4727]: I0929 10:36:39.475792 4727 generic.go:334] "Generic (PLEG): container finished" podID="fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2" containerID="36d067fbffcceb7e0d4f681e341950b2adaa1be1383d758c11abdc8f9dafa557" exitCode=0 Sep 29 10:36:39 crc kubenswrapper[4727]: I0929 10:36:39.475865 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerDied","Data":"36d067fbffcceb7e0d4f681e341950b2adaa1be1383d758c11abdc8f9dafa557"} Sep 29 10:36:40 crc kubenswrapper[4727]: I0929 10:36:40.482655 4727 generic.go:334] "Generic (PLEG): container finished" podID="fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2" containerID="d496a6af1d073ba76d9c0af2adb741a77db3cbddc6cbb7b8fbfdfd6f14b27c48" exitCode=0 Sep 29 10:36:40 crc kubenswrapper[4727]: I0929 10:36:40.482716 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerDied","Data":"d496a6af1d073ba76d9c0af2adb741a77db3cbddc6cbb7b8fbfdfd6f14b27c48"} Sep 29 10:36:41 crc kubenswrapper[4727]: I0929 10:36:41.491782 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"5170f21597afbcc29ab07e0daae3c78a90f9c0d95a5aa42e51da01ab5b458bca"} Sep 29 10:36:41 crc kubenswrapper[4727]: I0929 10:36:41.492452 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"92d646ad62b605ff5c1e1b4b8df67398e7893569486003c74e6f1b866ca440cc"} Sep 29 10:36:41 crc kubenswrapper[4727]: I0929 10:36:41.492467 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"df4b421bf1c29d7df4b64a8624b5e6b8ec6bffeecee369274c03efd239baeeed"} Sep 29 10:36:41 crc kubenswrapper[4727]: I0929 10:36:41.492476 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"7f2aedd57a5bc17a1414fb968f31ddf91405c77e9c74ab68bad5e203af9999ee"} Sep 29 10:36:42 crc kubenswrapper[4727]: I0929 10:36:42.500533 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"dc46a6b8e196567e83705ef071674c89e68bd66691c042b91e277848f74f7236"} Sep 29 10:36:43 crc kubenswrapper[4727]: I0929 10:36:43.510003 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jrlvl" event={"ID":"fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2","Type":"ContainerStarted","Data":"884f691ef528af221e6b5ee3337212cadfc81a4223a7f919124ea1ed4a5f6597"} Sep 29 10:36:43 crc kubenswrapper[4727]: I0929 10:36:43.511526 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:43 crc kubenswrapper[4727]: I0929 10:36:43.539176 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jrlvl" podStartSLOduration=7.244051598 podStartE2EDuration="15.53914605s" podCreationTimestamp="2025-09-29 10:36:28 +0000 UTC" firstStartedPulling="2025-09-29 10:36:29.590385394 +0000 UTC m=+859.763698786" lastFinishedPulling="2025-09-29 10:36:37.885479876 +0000 UTC m=+868.058793238" observedRunningTime="2025-09-29 10:36:43.537428406 +0000 UTC m=+873.710741788" watchObservedRunningTime="2025-09-29 10:36:43.53914605 +0000 UTC m=+873.712459412" Sep 29 10:36:44 crc kubenswrapper[4727]: I0929 10:36:44.062687 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:44 crc kubenswrapper[4727]: I0929 10:36:44.112789 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:49 crc kubenswrapper[4727]: I0929 10:36:49.072278 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-6pmt2" Sep 29 10:36:49 crc kubenswrapper[4727]: I0929 10:36:49.168655 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-b8fqx" Sep 29 10:36:50 crc kubenswrapper[4727]: I0929 10:36:50.640220 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-vh9j9" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.479700 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.480761 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.483508 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.483597 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.483666 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-j8rb8" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.507327 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.597185 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d7bl\" (UniqueName: \"kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl\") pod \"openstack-operator-index-s8wtj\" (UID: \"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9\") " pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.699756 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d7bl\" (UniqueName: \"kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl\") pod \"openstack-operator-index-s8wtj\" (UID: \"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9\") " pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.726956 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d7bl\" (UniqueName: \"kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl\") pod \"openstack-operator-index-s8wtj\" (UID: \"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9\") " pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:53 crc kubenswrapper[4727]: I0929 10:36:53.806753 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:54 crc kubenswrapper[4727]: I0929 10:36:54.217478 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:54 crc kubenswrapper[4727]: W0929 10:36:54.228357 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d0e5f0f_6188_42c2_a20b_6dd8fb030ae9.slice/crio-91f509c4863a6167425376c2cc31af84a6ec3eb0a4788d10532e8b95f1375375 WatchSource:0}: Error finding container 91f509c4863a6167425376c2cc31af84a6ec3eb0a4788d10532e8b95f1375375: Status 404 returned error can't find the container with id 91f509c4863a6167425376c2cc31af84a6ec3eb0a4788d10532e8b95f1375375 Sep 29 10:36:54 crc kubenswrapper[4727]: I0929 10:36:54.588483 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s8wtj" event={"ID":"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9","Type":"ContainerStarted","Data":"91f509c4863a6167425376c2cc31af84a6ec3eb0a4788d10532e8b95f1375375"} Sep 29 10:36:56 crc kubenswrapper[4727]: I0929 10:36:56.660360 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.257827 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8ps6m"] Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.259092 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.273041 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ps6m"] Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.360608 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rvb4\" (UniqueName: \"kubernetes.io/projected/9d7c10ba-9139-47d6-a681-507681066bff-kube-api-access-9rvb4\") pod \"openstack-operator-index-8ps6m\" (UID: \"9d7c10ba-9139-47d6-a681-507681066bff\") " pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.462871 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rvb4\" (UniqueName: \"kubernetes.io/projected/9d7c10ba-9139-47d6-a681-507681066bff-kube-api-access-9rvb4\") pod \"openstack-operator-index-8ps6m\" (UID: \"9d7c10ba-9139-47d6-a681-507681066bff\") " pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.490995 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rvb4\" (UniqueName: \"kubernetes.io/projected/9d7c10ba-9139-47d6-a681-507681066bff-kube-api-access-9rvb4\") pod \"openstack-operator-index-8ps6m\" (UID: \"9d7c10ba-9139-47d6-a681-507681066bff\") " pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.574772 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.612124 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s8wtj" event={"ID":"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9","Type":"ContainerStarted","Data":"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7"} Sep 29 10:36:57 crc kubenswrapper[4727]: I0929 10:36:57.612286 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-s8wtj" podUID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" containerName="registry-server" containerID="cri-o://0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7" gracePeriod=2 Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.026257 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-s8wtj" podStartSLOduration=2.430288803 podStartE2EDuration="5.026228619s" podCreationTimestamp="2025-09-29 10:36:53 +0000 UTC" firstStartedPulling="2025-09-29 10:36:54.231355386 +0000 UTC m=+884.404668748" lastFinishedPulling="2025-09-29 10:36:56.827295202 +0000 UTC m=+887.000608564" observedRunningTime="2025-09-29 10:36:57.634401138 +0000 UTC m=+887.807714520" watchObservedRunningTime="2025-09-29 10:36:58.026228619 +0000 UTC m=+888.199541981" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.028745 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.029484 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ps6m"] Sep 29 10:36:58 crc kubenswrapper[4727]: W0929 10:36:58.031620 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d7c10ba_9139_47d6_a681_507681066bff.slice/crio-15e82ff7981356b202aa68d9eeb63f531b0dc367d5821f1db36549a71fa0c9ca WatchSource:0}: Error finding container 15e82ff7981356b202aa68d9eeb63f531b0dc367d5821f1db36549a71fa0c9ca: Status 404 returned error can't find the container with id 15e82ff7981356b202aa68d9eeb63f531b0dc367d5821f1db36549a71fa0c9ca Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.170727 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d7bl\" (UniqueName: \"kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl\") pod \"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9\" (UID: \"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9\") " Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.180967 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl" (OuterVolumeSpecName: "kube-api-access-2d7bl") pod "1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" (UID: "1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9"). InnerVolumeSpecName "kube-api-access-2d7bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.271890 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d7bl\" (UniqueName: \"kubernetes.io/projected/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9-kube-api-access-2d7bl\") on node \"crc\" DevicePath \"\"" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.620893 4727 generic.go:334] "Generic (PLEG): container finished" podID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" containerID="0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7" exitCode=0 Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.620983 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s8wtj" event={"ID":"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9","Type":"ContainerDied","Data":"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7"} Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.621021 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s8wtj" event={"ID":"1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9","Type":"ContainerDied","Data":"91f509c4863a6167425376c2cc31af84a6ec3eb0a4788d10532e8b95f1375375"} Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.621044 4727 scope.go:117] "RemoveContainer" containerID="0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.621193 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s8wtj" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.625395 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ps6m" event={"ID":"9d7c10ba-9139-47d6-a681-507681066bff","Type":"ContainerStarted","Data":"916205caa48783357f215d4029dbd3d882dab91f1b931bc3be233fc9eae3098a"} Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.625460 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ps6m" event={"ID":"9d7c10ba-9139-47d6-a681-507681066bff","Type":"ContainerStarted","Data":"15e82ff7981356b202aa68d9eeb63f531b0dc367d5821f1db36549a71fa0c9ca"} Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.642717 4727 scope.go:117] "RemoveContainer" containerID="0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7" Sep 29 10:36:58 crc kubenswrapper[4727]: E0929 10:36:58.643284 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7\": container with ID starting with 0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7 not found: ID does not exist" containerID="0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.643390 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7"} err="failed to get container status \"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7\": rpc error: code = NotFound desc = could not find container \"0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7\": container with ID starting with 0d1ef382f562c5711706f2e0a0a701f22575255ad60b2bad81dbc539c96221b7 not found: ID does not exist" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.650282 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8ps6m" podStartSLOduration=1.590436461 podStartE2EDuration="1.650264271s" podCreationTimestamp="2025-09-29 10:36:57 +0000 UTC" firstStartedPulling="2025-09-29 10:36:58.035902289 +0000 UTC m=+888.209215651" lastFinishedPulling="2025-09-29 10:36:58.095730099 +0000 UTC m=+888.269043461" observedRunningTime="2025-09-29 10:36:58.649809539 +0000 UTC m=+888.823122931" watchObservedRunningTime="2025-09-29 10:36:58.650264271 +0000 UTC m=+888.823577633" Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.664634 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:58 crc kubenswrapper[4727]: I0929 10:36:58.667464 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-s8wtj"] Sep 29 10:36:59 crc kubenswrapper[4727]: I0929 10:36:59.066872 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jrlvl" Sep 29 10:36:59 crc kubenswrapper[4727]: I0929 10:36:59.116542 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" path="/var/lib/kubelet/pods/1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9/volumes" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.669389 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:02 crc kubenswrapper[4727]: E0929 10:37:02.673001 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" containerName="registry-server" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.673111 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" containerName="registry-server" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.673406 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d0e5f0f-6188-42c2-a20b-6dd8fb030ae9" containerName="registry-server" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.674569 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.682457 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.842175 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k796w\" (UniqueName: \"kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.842884 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.842927 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.944579 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.944659 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.944709 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k796w\" (UniqueName: \"kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.946021 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.946036 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:02 crc kubenswrapper[4727]: I0929 10:37:02.969376 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k796w\" (UniqueName: \"kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w\") pod \"community-operators-6s6n5\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:03 crc kubenswrapper[4727]: I0929 10:37:03.002762 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:03 crc kubenswrapper[4727]: I0929 10:37:03.568057 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:03 crc kubenswrapper[4727]: W0929 10:37:03.579621 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa95b00a_4e2a_4530_8a31_4e4a944c95ca.slice/crio-c432c9259718d8a24e34c35a946499b55681e4716fb7cd1ee8a7aa747801ea08 WatchSource:0}: Error finding container c432c9259718d8a24e34c35a946499b55681e4716fb7cd1ee8a7aa747801ea08: Status 404 returned error can't find the container with id c432c9259718d8a24e34c35a946499b55681e4716fb7cd1ee8a7aa747801ea08 Sep 29 10:37:03 crc kubenswrapper[4727]: I0929 10:37:03.662807 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerStarted","Data":"c432c9259718d8a24e34c35a946499b55681e4716fb7cd1ee8a7aa747801ea08"} Sep 29 10:37:04 crc kubenswrapper[4727]: I0929 10:37:04.674763 4727 generic.go:334] "Generic (PLEG): container finished" podID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerID="22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b" exitCode=0 Sep 29 10:37:04 crc kubenswrapper[4727]: I0929 10:37:04.674896 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerDied","Data":"22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b"} Sep 29 10:37:05 crc kubenswrapper[4727]: I0929 10:37:05.686088 4727 generic.go:334] "Generic (PLEG): container finished" podID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerID="ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125" exitCode=0 Sep 29 10:37:05 crc kubenswrapper[4727]: I0929 10:37:05.686199 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerDied","Data":"ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125"} Sep 29 10:37:06 crc kubenswrapper[4727]: I0929 10:37:06.712964 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerStarted","Data":"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0"} Sep 29 10:37:06 crc kubenswrapper[4727]: I0929 10:37:06.734555 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6s6n5" podStartSLOduration=3.302205705 podStartE2EDuration="4.734478772s" podCreationTimestamp="2025-09-29 10:37:02 +0000 UTC" firstStartedPulling="2025-09-29 10:37:04.677666001 +0000 UTC m=+894.850979363" lastFinishedPulling="2025-09-29 10:37:06.109939068 +0000 UTC m=+896.283252430" observedRunningTime="2025-09-29 10:37:06.731923077 +0000 UTC m=+896.905236449" watchObservedRunningTime="2025-09-29 10:37:06.734478772 +0000 UTC m=+896.907792134" Sep 29 10:37:07 crc kubenswrapper[4727]: I0929 10:37:07.575573 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:37:07 crc kubenswrapper[4727]: I0929 10:37:07.576307 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:37:07 crc kubenswrapper[4727]: I0929 10:37:07.608080 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:37:07 crc kubenswrapper[4727]: I0929 10:37:07.749890 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-8ps6m" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.717619 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8"] Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.719038 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.722425 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dfmjz" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.742761 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8"] Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.849974 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.850080 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.850130 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prsr7\" (UniqueName: \"kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.951066 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.951127 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.951169 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prsr7\" (UniqueName: \"kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.952277 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.952740 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:09 crc kubenswrapper[4727]: I0929 10:37:09.988557 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prsr7\" (UniqueName: \"kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7\") pod \"3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:10 crc kubenswrapper[4727]: I0929 10:37:10.046322 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:10 crc kubenswrapper[4727]: I0929 10:37:10.506802 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8"] Sep 29 10:37:10 crc kubenswrapper[4727]: I0929 10:37:10.745137 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" event={"ID":"9968f970-53fe-440d-9e00-2f3409f8201c","Type":"ContainerStarted","Data":"de9052895fb70e4e6fd5eba29b42d75f765d1d883b35817e1dbdcf9184de92d6"} Sep 29 10:37:12 crc kubenswrapper[4727]: I0929 10:37:12.768647 4727 generic.go:334] "Generic (PLEG): container finished" podID="9968f970-53fe-440d-9e00-2f3409f8201c" containerID="a03ab65f4af049585f6232f4442abc6a08a8f7038bb662c336d6db43e8ae7156" exitCode=0 Sep 29 10:37:12 crc kubenswrapper[4727]: I0929 10:37:12.768733 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" event={"ID":"9968f970-53fe-440d-9e00-2f3409f8201c","Type":"ContainerDied","Data":"a03ab65f4af049585f6232f4442abc6a08a8f7038bb662c336d6db43e8ae7156"} Sep 29 10:37:13 crc kubenswrapper[4727]: I0929 10:37:13.003733 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:13 crc kubenswrapper[4727]: I0929 10:37:13.003901 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:13 crc kubenswrapper[4727]: I0929 10:37:13.065299 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:13 crc kubenswrapper[4727]: I0929 10:37:13.823280 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:14 crc kubenswrapper[4727]: I0929 10:37:14.790108 4727 generic.go:334] "Generic (PLEG): container finished" podID="9968f970-53fe-440d-9e00-2f3409f8201c" containerID="9b4b465ec8092c8dc4199182fb58a287e51c6130acc35b0b4ecbea76e9c20d03" exitCode=0 Sep 29 10:37:14 crc kubenswrapper[4727]: I0929 10:37:14.790209 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" event={"ID":"9968f970-53fe-440d-9e00-2f3409f8201c","Type":"ContainerDied","Data":"9b4b465ec8092c8dc4199182fb58a287e51c6130acc35b0b4ecbea76e9c20d03"} Sep 29 10:37:15 crc kubenswrapper[4727]: I0929 10:37:15.797713 4727 generic.go:334] "Generic (PLEG): container finished" podID="9968f970-53fe-440d-9e00-2f3409f8201c" containerID="d9b3368bca570c2f31771daf7d4dc6f6630264b21bf51b10bfeb56d4cf364324" exitCode=0 Sep 29 10:37:15 crc kubenswrapper[4727]: I0929 10:37:15.797816 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" event={"ID":"9968f970-53fe-440d-9e00-2f3409f8201c","Type":"ContainerDied","Data":"d9b3368bca570c2f31771daf7d4dc6f6630264b21bf51b10bfeb56d4cf364324"} Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.045894 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.084331 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:17 crc kubenswrapper[4727]: E0929 10:37:17.087867 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="pull" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.087896 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="pull" Sep 29 10:37:17 crc kubenswrapper[4727]: E0929 10:37:17.087991 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="extract" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.088020 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="extract" Sep 29 10:37:17 crc kubenswrapper[4727]: E0929 10:37:17.088047 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="util" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.088055 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="util" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.088496 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9968f970-53fe-440d-9e00-2f3409f8201c" containerName="extract" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.090558 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.100837 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157097 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prsr7\" (UniqueName: \"kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7\") pod \"9968f970-53fe-440d-9e00-2f3409f8201c\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157209 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle\") pod \"9968f970-53fe-440d-9e00-2f3409f8201c\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157265 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util\") pod \"9968f970-53fe-440d-9e00-2f3409f8201c\" (UID: \"9968f970-53fe-440d-9e00-2f3409f8201c\") " Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157468 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157495 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98mrr\" (UniqueName: \"kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.157545 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.160543 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle" (OuterVolumeSpecName: "bundle") pod "9968f970-53fe-440d-9e00-2f3409f8201c" (UID: "9968f970-53fe-440d-9e00-2f3409f8201c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.162818 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7" (OuterVolumeSpecName: "kube-api-access-prsr7") pod "9968f970-53fe-440d-9e00-2f3409f8201c" (UID: "9968f970-53fe-440d-9e00-2f3409f8201c"). InnerVolumeSpecName "kube-api-access-prsr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.255797 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.256539 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6s6n5" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="registry-server" containerID="cri-o://368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0" gracePeriod=2 Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.258751 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.258923 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98mrr\" (UniqueName: \"kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.259097 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.259277 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prsr7\" (UniqueName: \"kubernetes.io/projected/9968f970-53fe-440d-9e00-2f3409f8201c-kube-api-access-prsr7\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.259442 4727 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.260064 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.260427 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.283372 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98mrr\" (UniqueName: \"kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr\") pod \"redhat-marketplace-fl45b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.419655 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.814748 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" event={"ID":"9968f970-53fe-440d-9e00-2f3409f8201c","Type":"ContainerDied","Data":"de9052895fb70e4e6fd5eba29b42d75f765d1d883b35817e1dbdcf9184de92d6"} Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.814789 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de9052895fb70e4e6fd5eba29b42d75f765d1d883b35817e1dbdcf9184de92d6" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.815277 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8" Sep 29 10:37:17 crc kubenswrapper[4727]: I0929 10:37:17.831822 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:17 crc kubenswrapper[4727]: W0929 10:37:17.841478 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2d075b2_8ca4_4a56_bb35_42b7c3004a8b.slice/crio-66c1327d99e2f3a131df6f2641cc69a16a45df8dab95450a98cb58c50b8d5150 WatchSource:0}: Error finding container 66c1327d99e2f3a131df6f2641cc69a16a45df8dab95450a98cb58c50b8d5150: Status 404 returned error can't find the container with id 66c1327d99e2f3a131df6f2641cc69a16a45df8dab95450a98cb58c50b8d5150 Sep 29 10:37:18 crc kubenswrapper[4727]: I0929 10:37:18.818369 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util" (OuterVolumeSpecName: "util") pod "9968f970-53fe-440d-9e00-2f3409f8201c" (UID: "9968f970-53fe-440d-9e00-2f3409f8201c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:18 crc kubenswrapper[4727]: I0929 10:37:18.822543 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerStarted","Data":"66c1327d99e2f3a131df6f2641cc69a16a45df8dab95450a98cb58c50b8d5150"} Sep 29 10:37:18 crc kubenswrapper[4727]: I0929 10:37:18.879794 4727 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9968f970-53fe-440d-9e00-2f3409f8201c-util\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.246940 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.247027 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.704295 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.791156 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities\") pod \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.791222 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content\") pod \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.791321 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k796w\" (UniqueName: \"kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w\") pod \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\" (UID: \"aa95b00a-4e2a-4530-8a31-4e4a944c95ca\") " Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.792681 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities" (OuterVolumeSpecName: "utilities") pod "aa95b00a-4e2a-4530-8a31-4e4a944c95ca" (UID: "aa95b00a-4e2a-4530-8a31-4e4a944c95ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.796515 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w" (OuterVolumeSpecName: "kube-api-access-k796w") pod "aa95b00a-4e2a-4530-8a31-4e4a944c95ca" (UID: "aa95b00a-4e2a-4530-8a31-4e4a944c95ca"). InnerVolumeSpecName "kube-api-access-k796w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.834129 4727 generic.go:334] "Generic (PLEG): container finished" podID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerID="4d314e8078910520d1e3e383be26693a71a49be2f78e54e22f9ff4c8d1fce34e" exitCode=0 Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.834199 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerDied","Data":"4d314e8078910520d1e3e383be26693a71a49be2f78e54e22f9ff4c8d1fce34e"} Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.840818 4727 generic.go:334] "Generic (PLEG): container finished" podID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerID="368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0" exitCode=0 Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.840851 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerDied","Data":"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0"} Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.840896 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6s6n5" event={"ID":"aa95b00a-4e2a-4530-8a31-4e4a944c95ca","Type":"ContainerDied","Data":"c432c9259718d8a24e34c35a946499b55681e4716fb7cd1ee8a7aa747801ea08"} Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.840912 4727 scope.go:117] "RemoveContainer" containerID="368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.841048 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6s6n5" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.864981 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa95b00a-4e2a-4530-8a31-4e4a944c95ca" (UID: "aa95b00a-4e2a-4530-8a31-4e4a944c95ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.878193 4727 scope.go:117] "RemoveContainer" containerID="ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.894397 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.894428 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.894439 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k796w\" (UniqueName: \"kubernetes.io/projected/aa95b00a-4e2a-4530-8a31-4e4a944c95ca-kube-api-access-k796w\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.898732 4727 scope.go:117] "RemoveContainer" containerID="22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.938905 4727 scope.go:117] "RemoveContainer" containerID="368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0" Sep 29 10:37:19 crc kubenswrapper[4727]: E0929 10:37:19.941727 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0\": container with ID starting with 368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0 not found: ID does not exist" containerID="368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.941772 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0"} err="failed to get container status \"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0\": rpc error: code = NotFound desc = could not find container \"368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0\": container with ID starting with 368d893e8f01b495a698260a9ec89fb4faf27270fb6a88472cc88f39b137dec0 not found: ID does not exist" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.941801 4727 scope.go:117] "RemoveContainer" containerID="ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125" Sep 29 10:37:19 crc kubenswrapper[4727]: E0929 10:37:19.942101 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125\": container with ID starting with ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125 not found: ID does not exist" containerID="ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.942147 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125"} err="failed to get container status \"ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125\": rpc error: code = NotFound desc = could not find container \"ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125\": container with ID starting with ba5117a6d160da47f4be104b0a6c4d9b4500c72dc600ee0f83a968a71cafd125 not found: ID does not exist" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.942175 4727 scope.go:117] "RemoveContainer" containerID="22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b" Sep 29 10:37:19 crc kubenswrapper[4727]: E0929 10:37:19.942941 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b\": container with ID starting with 22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b not found: ID does not exist" containerID="22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b" Sep 29 10:37:19 crc kubenswrapper[4727]: I0929 10:37:19.942975 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b"} err="failed to get container status \"22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b\": rpc error: code = NotFound desc = could not find container \"22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b\": container with ID starting with 22355802d499a1a2a8dcc40307ab2eb274cead0b8352472b9596bf602b3f194b not found: ID does not exist" Sep 29 10:37:20 crc kubenswrapper[4727]: I0929 10:37:20.168754 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:20 crc kubenswrapper[4727]: I0929 10:37:20.174265 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6s6n5"] Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.117759 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" path="/var/lib/kubelet/pods/aa95b00a-4e2a-4530-8a31-4e4a944c95ca/volumes" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.656057 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62"] Sep 29 10:37:21 crc kubenswrapper[4727]: E0929 10:37:21.656602 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="extract-utilities" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.656613 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="extract-utilities" Sep 29 10:37:21 crc kubenswrapper[4727]: E0929 10:37:21.656625 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="extract-content" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.656632 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="extract-content" Sep 29 10:37:21 crc kubenswrapper[4727]: E0929 10:37:21.656641 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="registry-server" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.656647 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="registry-server" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.656758 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa95b00a-4e2a-4530-8a31-4e4a944c95ca" containerName="registry-server" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.657429 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.665907 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-g8ljd" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.698111 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62"] Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.717724 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bglsr\" (UniqueName: \"kubernetes.io/projected/b2fb3066-6094-4799-b68d-67ab396b160e-kube-api-access-bglsr\") pod \"openstack-operator-controller-operator-85bb87788c-jvd62\" (UID: \"b2fb3066-6094-4799-b68d-67ab396b160e\") " pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.818706 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bglsr\" (UniqueName: \"kubernetes.io/projected/b2fb3066-6094-4799-b68d-67ab396b160e-kube-api-access-bglsr\") pod \"openstack-operator-controller-operator-85bb87788c-jvd62\" (UID: \"b2fb3066-6094-4799-b68d-67ab396b160e\") " pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.845867 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bglsr\" (UniqueName: \"kubernetes.io/projected/b2fb3066-6094-4799-b68d-67ab396b160e-kube-api-access-bglsr\") pod \"openstack-operator-controller-operator-85bb87788c-jvd62\" (UID: \"b2fb3066-6094-4799-b68d-67ab396b160e\") " pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.861346 4727 generic.go:334] "Generic (PLEG): container finished" podID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerID="b4fca4e1d7a05420c40416d37b948f0e171dcad5bb999eafc887bc2bcb1580c1" exitCode=0 Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.861426 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerDied","Data":"b4fca4e1d7a05420c40416d37b948f0e171dcad5bb999eafc887bc2bcb1580c1"} Sep 29 10:37:21 crc kubenswrapper[4727]: I0929 10:37:21.973440 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:22 crc kubenswrapper[4727]: I0929 10:37:22.396280 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62"] Sep 29 10:37:22 crc kubenswrapper[4727]: W0929 10:37:22.403430 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2fb3066_6094_4799_b68d_67ab396b160e.slice/crio-a3e6049b7555562e128adaa7d5a6a7e75ef04f986b586b57bef493004fde81ea WatchSource:0}: Error finding container a3e6049b7555562e128adaa7d5a6a7e75ef04f986b586b57bef493004fde81ea: Status 404 returned error can't find the container with id a3e6049b7555562e128adaa7d5a6a7e75ef04f986b586b57bef493004fde81ea Sep 29 10:37:22 crc kubenswrapper[4727]: I0929 10:37:22.868001 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" event={"ID":"b2fb3066-6094-4799-b68d-67ab396b160e","Type":"ContainerStarted","Data":"a3e6049b7555562e128adaa7d5a6a7e75ef04f986b586b57bef493004fde81ea"} Sep 29 10:37:22 crc kubenswrapper[4727]: I0929 10:37:22.870477 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerStarted","Data":"ba5287243e5f026695dd01c4e87451ec854c2af09671b64420fe6f57dbdce05a"} Sep 29 10:37:22 crc kubenswrapper[4727]: I0929 10:37:22.899388 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fl45b" podStartSLOduration=3.121728077 podStartE2EDuration="5.899363122s" podCreationTimestamp="2025-09-29 10:37:17 +0000 UTC" firstStartedPulling="2025-09-29 10:37:19.835795133 +0000 UTC m=+910.009108495" lastFinishedPulling="2025-09-29 10:37:22.613430178 +0000 UTC m=+912.786743540" observedRunningTime="2025-09-29 10:37:22.893981613 +0000 UTC m=+913.067294975" watchObservedRunningTime="2025-09-29 10:37:22.899363122 +0000 UTC m=+913.072676484" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.420668 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.420975 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.464853 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.686848 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.688382 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.701306 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.703601 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.703635 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwwq8\" (UniqueName: \"kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.703678 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.804746 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.804842 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.804868 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwwq8\" (UniqueName: \"kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.805200 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.805568 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.826936 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwwq8\" (UniqueName: \"kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8\") pod \"certified-operators-7r2hj\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:27 crc kubenswrapper[4727]: I0929 10:37:27.959702 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:28 crc kubenswrapper[4727]: I0929 10:37:28.008674 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:28 crc kubenswrapper[4727]: I0929 10:37:28.894538 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:28 crc kubenswrapper[4727]: I0929 10:37:28.915765 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerStarted","Data":"e6827187090076412ed3c2a5ca4ae0186a618ebf275593b44592e5f6379eb9be"} Sep 29 10:37:28 crc kubenswrapper[4727]: I0929 10:37:28.917935 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" event={"ID":"b2fb3066-6094-4799-b68d-67ab396b160e","Type":"ContainerStarted","Data":"69ea3d5e26a6aae2875372562ea3cce0229b17cd0873ff6b1d9f536d182466fd"} Sep 29 10:37:29 crc kubenswrapper[4727]: I0929 10:37:29.923384 4727 generic.go:334] "Generic (PLEG): container finished" podID="807456e2-2452-4416-b7d7-399bd51e3db5" containerID="77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452" exitCode=0 Sep 29 10:37:29 crc kubenswrapper[4727]: I0929 10:37:29.923497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerDied","Data":"77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452"} Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.053574 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.053832 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fl45b" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="registry-server" containerID="cri-o://ba5287243e5f026695dd01c4e87451ec854c2af09671b64420fe6f57dbdce05a" gracePeriod=2 Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.935750 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" event={"ID":"b2fb3066-6094-4799-b68d-67ab396b160e","Type":"ContainerStarted","Data":"98594099c1a34972fa63adb5dd565d15bd824f0794ee381a50708ef32b349c9f"} Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.936091 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.937485 4727 generic.go:334] "Generic (PLEG): container finished" podID="807456e2-2452-4416-b7d7-399bd51e3db5" containerID="39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece" exitCode=0 Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.937586 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerDied","Data":"39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece"} Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.947811 4727 generic.go:334] "Generic (PLEG): container finished" podID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerID="ba5287243e5f026695dd01c4e87451ec854c2af09671b64420fe6f57dbdce05a" exitCode=0 Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.947868 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerDied","Data":"ba5287243e5f026695dd01c4e87451ec854c2af09671b64420fe6f57dbdce05a"} Sep 29 10:37:31 crc kubenswrapper[4727]: I0929 10:37:31.990427 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" podStartSLOduration=2.670986564 podStartE2EDuration="10.990410892s" podCreationTimestamp="2025-09-29 10:37:21 +0000 UTC" firstStartedPulling="2025-09-29 10:37:22.405705458 +0000 UTC m=+912.579018820" lastFinishedPulling="2025-09-29 10:37:30.725129786 +0000 UTC m=+920.898443148" observedRunningTime="2025-09-29 10:37:31.985466535 +0000 UTC m=+922.158779897" watchObservedRunningTime="2025-09-29 10:37:31.990410892 +0000 UTC m=+922.163724254" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.101784 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.272647 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98mrr\" (UniqueName: \"kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr\") pod \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.272694 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content\") pod \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.272726 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities\") pod \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\" (UID: \"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b\") " Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.274701 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities" (OuterVolumeSpecName: "utilities") pod "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" (UID: "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.287502 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr" (OuterVolumeSpecName: "kube-api-access-98mrr") pod "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" (UID: "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b"). InnerVolumeSpecName "kube-api-access-98mrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.288362 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" (UID: "b2d075b2-8ca4-4a56-bb35-42b7c3004a8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.373975 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98mrr\" (UniqueName: \"kubernetes.io/projected/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-kube-api-access-98mrr\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.374021 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.374030 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.955776 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fl45b" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.955909 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fl45b" event={"ID":"b2d075b2-8ca4-4a56-bb35-42b7c3004a8b","Type":"ContainerDied","Data":"66c1327d99e2f3a131df6f2641cc69a16a45df8dab95450a98cb58c50b8d5150"} Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.957532 4727 scope.go:117] "RemoveContainer" containerID="ba5287243e5f026695dd01c4e87451ec854c2af09671b64420fe6f57dbdce05a" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.972260 4727 scope.go:117] "RemoveContainer" containerID="b4fca4e1d7a05420c40416d37b948f0e171dcad5bb999eafc887bc2bcb1580c1" Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.980355 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:32 crc kubenswrapper[4727]: I0929 10:37:32.984628 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fl45b"] Sep 29 10:37:33 crc kubenswrapper[4727]: I0929 10:37:33.002850 4727 scope.go:117] "RemoveContainer" containerID="4d314e8078910520d1e3e383be26693a71a49be2f78e54e22f9ff4c8d1fce34e" Sep 29 10:37:33 crc kubenswrapper[4727]: I0929 10:37:33.116460 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" path="/var/lib/kubelet/pods/b2d075b2-8ca4-4a56-bb35-42b7c3004a8b/volumes" Sep 29 10:37:33 crc kubenswrapper[4727]: I0929 10:37:33.962229 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerStarted","Data":"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6"} Sep 29 10:37:33 crc kubenswrapper[4727]: I0929 10:37:33.987205 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7r2hj" podStartSLOduration=4.09181137 podStartE2EDuration="6.987181737s" podCreationTimestamp="2025-09-29 10:37:27 +0000 UTC" firstStartedPulling="2025-09-29 10:37:29.944727438 +0000 UTC m=+920.118040800" lastFinishedPulling="2025-09-29 10:37:32.840097805 +0000 UTC m=+923.013411167" observedRunningTime="2025-09-29 10:37:33.982030514 +0000 UTC m=+924.155343886" watchObservedRunningTime="2025-09-29 10:37:33.987181737 +0000 UTC m=+924.160495099" Sep 29 10:37:38 crc kubenswrapper[4727]: I0929 10:37:38.009536 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:38 crc kubenswrapper[4727]: I0929 10:37:38.009930 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:38 crc kubenswrapper[4727]: I0929 10:37:38.059247 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:39 crc kubenswrapper[4727]: I0929 10:37:39.033142 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:40 crc kubenswrapper[4727]: I0929 10:37:40.053156 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:40 crc kubenswrapper[4727]: I0929 10:37:40.999664 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7r2hj" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="registry-server" containerID="cri-o://03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6" gracePeriod=2 Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.357126 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.496921 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities\") pod \"807456e2-2452-4416-b7d7-399bd51e3db5\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.497376 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content\") pod \"807456e2-2452-4416-b7d7-399bd51e3db5\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.497594 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwwq8\" (UniqueName: \"kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8\") pod \"807456e2-2452-4416-b7d7-399bd51e3db5\" (UID: \"807456e2-2452-4416-b7d7-399bd51e3db5\") " Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.497812 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities" (OuterVolumeSpecName: "utilities") pod "807456e2-2452-4416-b7d7-399bd51e3db5" (UID: "807456e2-2452-4416-b7d7-399bd51e3db5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.497903 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.507047 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8" (OuterVolumeSpecName: "kube-api-access-wwwq8") pod "807456e2-2452-4416-b7d7-399bd51e3db5" (UID: "807456e2-2452-4416-b7d7-399bd51e3db5"). InnerVolumeSpecName "kube-api-access-wwwq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.563166 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "807456e2-2452-4416-b7d7-399bd51e3db5" (UID: "807456e2-2452-4416-b7d7-399bd51e3db5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.599250 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwwq8\" (UniqueName: \"kubernetes.io/projected/807456e2-2452-4416-b7d7-399bd51e3db5-kube-api-access-wwwq8\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.599289 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/807456e2-2452-4416-b7d7-399bd51e3db5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:37:41 crc kubenswrapper[4727]: I0929 10:37:41.976455 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-85bb87788c-jvd62" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.008019 4727 generic.go:334] "Generic (PLEG): container finished" podID="807456e2-2452-4416-b7d7-399bd51e3db5" containerID="03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6" exitCode=0 Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.008074 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7r2hj" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.008066 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerDied","Data":"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6"} Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.008144 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7r2hj" event={"ID":"807456e2-2452-4416-b7d7-399bd51e3db5","Type":"ContainerDied","Data":"e6827187090076412ed3c2a5ca4ae0186a618ebf275593b44592e5f6379eb9be"} Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.008167 4727 scope.go:117] "RemoveContainer" containerID="03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.027244 4727 scope.go:117] "RemoveContainer" containerID="39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.039757 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.049934 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7r2hj"] Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.069837 4727 scope.go:117] "RemoveContainer" containerID="77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.087451 4727 scope.go:117] "RemoveContainer" containerID="03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6" Sep 29 10:37:42 crc kubenswrapper[4727]: E0929 10:37:42.087909 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6\": container with ID starting with 03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6 not found: ID does not exist" containerID="03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.087949 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6"} err="failed to get container status \"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6\": rpc error: code = NotFound desc = could not find container \"03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6\": container with ID starting with 03a83d6a9d930962eaf9f5f6dc2cc2113f73c2ed3136425af41d31287ea63dd6 not found: ID does not exist" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.087974 4727 scope.go:117] "RemoveContainer" containerID="39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece" Sep 29 10:37:42 crc kubenswrapper[4727]: E0929 10:37:42.088245 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece\": container with ID starting with 39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece not found: ID does not exist" containerID="39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.088329 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece"} err="failed to get container status \"39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece\": rpc error: code = NotFound desc = could not find container \"39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece\": container with ID starting with 39f7fefa8e13a74cb1a1773c11a2483700cd54053468bd29e1d99c8d8f801ece not found: ID does not exist" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.088445 4727 scope.go:117] "RemoveContainer" containerID="77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452" Sep 29 10:37:42 crc kubenswrapper[4727]: E0929 10:37:42.088729 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452\": container with ID starting with 77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452 not found: ID does not exist" containerID="77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452" Sep 29 10:37:42 crc kubenswrapper[4727]: I0929 10:37:42.088756 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452"} err="failed to get container status \"77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452\": rpc error: code = NotFound desc = could not find container \"77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452\": container with ID starting with 77607d781ec51bdc13325a64f4cdb8daef71b6e30626e6f66cdb5bff6eba4452 not found: ID does not exist" Sep 29 10:37:43 crc kubenswrapper[4727]: I0929 10:37:43.115960 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" path="/var/lib/kubelet/pods/807456e2-2452-4416-b7d7-399bd51e3db5/volumes" Sep 29 10:37:49 crc kubenswrapper[4727]: I0929 10:37:49.247310 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:37:49 crc kubenswrapper[4727]: I0929 10:37:49.247910 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.759683 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb"] Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760445 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="extract-utilities" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760457 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="extract-utilities" Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760473 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="extract-content" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760479 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="extract-content" Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760487 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="extract-utilities" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760493 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="extract-utilities" Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760503 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="extract-content" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760509 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="extract-content" Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760522 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760528 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: E0929 10:37:57.760535 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760540 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760661 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="807456e2-2452-4416-b7d7-399bd51e3db5" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.760677 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d075b2-8ca4-4a56-bb35-42b7c3004a8b" containerName="registry-server" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.761289 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.766166 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-4lwlk" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.770479 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.771422 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.773528 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-786sh" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.776197 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.793679 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.794881 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.800095 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.806290 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nckxd" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.818108 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.821884 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.824983 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.825807 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p9qv6" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.849580 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.850743 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.852908 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-s8z4c" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.857405 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.872447 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.873479 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.878195 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-f8sc9" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.896417 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.916278 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.932017 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tp5\" (UniqueName: \"kubernetes.io/projected/44a5b923-82b5-4af1-ad00-ba65998598ff-kube-api-access-95tp5\") pod \"barbican-operator-controller-manager-6495d75b5-rbtdb\" (UID: \"44a5b923-82b5-4af1-ad00-ba65998598ff\") " pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.932314 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7hwt\" (UniqueName: \"kubernetes.io/projected/2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609-kube-api-access-z7hwt\") pod \"designate-operator-controller-manager-7d74f4d695-8txnk\" (UID: \"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609\") " pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.932604 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52pc9\" (UniqueName: \"kubernetes.io/projected/2cb520ac-7d2f-46ab-b8bc-5f515485d278-kube-api-access-52pc9\") pod \"glance-operator-controller-manager-67b5d44b7f-vx2tw\" (UID: \"2cb520ac-7d2f-46ab-b8bc-5f515485d278\") " pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.932710 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fbmc\" (UniqueName: \"kubernetes.io/projected/f6f41603-6f22-4086-be87-c1a0062a691d-kube-api-access-8fbmc\") pod \"cinder-operator-controller-manager-748c574d75-8s5z5\" (UID: \"f6f41603-6f22-4086-be87-c1a0062a691d\") " pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.945893 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z"] Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.948240 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.953074 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hlvjj" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.953805 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 29 10:37:57 crc kubenswrapper[4727]: I0929 10:37:57.991833 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.008061 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.009192 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.013748 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-z26dm" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.018879 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.029637 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036544 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbgl5\" (UniqueName: \"kubernetes.io/projected/aef52f91-e131-49a0-84c8-ba76db16d118-kube-api-access-pbgl5\") pod \"heat-operator-controller-manager-8ff95898-6b6gt\" (UID: \"aef52f91-e131-49a0-84c8-ba76db16d118\") " pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036619 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036651 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52pc9\" (UniqueName: \"kubernetes.io/projected/2cb520ac-7d2f-46ab-b8bc-5f515485d278-kube-api-access-52pc9\") pod \"glance-operator-controller-manager-67b5d44b7f-vx2tw\" (UID: \"2cb520ac-7d2f-46ab-b8bc-5f515485d278\") " pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036667 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fbmc\" (UniqueName: \"kubernetes.io/projected/f6f41603-6f22-4086-be87-c1a0062a691d-kube-api-access-8fbmc\") pod \"cinder-operator-controller-manager-748c574d75-8s5z5\" (UID: \"f6f41603-6f22-4086-be87-c1a0062a691d\") " pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036716 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tp5\" (UniqueName: \"kubernetes.io/projected/44a5b923-82b5-4af1-ad00-ba65998598ff-kube-api-access-95tp5\") pod \"barbican-operator-controller-manager-6495d75b5-rbtdb\" (UID: \"44a5b923-82b5-4af1-ad00-ba65998598ff\") " pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036743 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7hwt\" (UniqueName: \"kubernetes.io/projected/2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609-kube-api-access-z7hwt\") pod \"designate-operator-controller-manager-7d74f4d695-8txnk\" (UID: \"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609\") " pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036762 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffv29\" (UniqueName: \"kubernetes.io/projected/66342226-bc85-46a0-8d84-0627c26b196c-kube-api-access-ffv29\") pod \"horizon-operator-controller-manager-695847bc78-fcs2w\" (UID: \"66342226-bc85-46a0-8d84-0627c26b196c\") " pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.036787 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmhvw\" (UniqueName: \"kubernetes.io/projected/e5545a7b-7c61-4e6b-92e9-3a22045a2133-kube-api-access-xmhvw\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.037819 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.041000 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-cns9k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.041213 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.055442 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.056882 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.064567 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52pc9\" (UniqueName: \"kubernetes.io/projected/2cb520ac-7d2f-46ab-b8bc-5f515485d278-kube-api-access-52pc9\") pod \"glance-operator-controller-manager-67b5d44b7f-vx2tw\" (UID: \"2cb520ac-7d2f-46ab-b8bc-5f515485d278\") " pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.064618 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7hwt\" (UniqueName: \"kubernetes.io/projected/2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609-kube-api-access-z7hwt\") pod \"designate-operator-controller-manager-7d74f4d695-8txnk\" (UID: \"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609\") " pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.067909 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tp5\" (UniqueName: \"kubernetes.io/projected/44a5b923-82b5-4af1-ad00-ba65998598ff-kube-api-access-95tp5\") pod \"barbican-operator-controller-manager-6495d75b5-rbtdb\" (UID: \"44a5b923-82b5-4af1-ad00-ba65998598ff\") " pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.068579 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fbmc\" (UniqueName: \"kubernetes.io/projected/f6f41603-6f22-4086-be87-c1a0062a691d-kube-api-access-8fbmc\") pod \"cinder-operator-controller-manager-748c574d75-8s5z5\" (UID: \"f6f41603-6f22-4086-be87-c1a0062a691d\") " pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.068723 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.073630 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.074970 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.081871 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-ncwp7" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.084151 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.085070 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6xbvw" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.094456 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.095862 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.095984 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.096656 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.103481 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-tpdzn" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.117254 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.122996 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.124695 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.127572 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-mbwcl" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.136360 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.140003 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.146858 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g9wn\" (UniqueName: \"kubernetes.io/projected/6f3ebbd5-f00d-4db3-8160-03e9dce230e4-kube-api-access-4g9wn\") pod \"keystone-operator-controller-manager-7bf498966c-vktwk\" (UID: \"6f3ebbd5-f00d-4db3-8160-03e9dce230e4\") " pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.146922 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.147001 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krpk5\" (UniqueName: \"kubernetes.io/projected/51327d2c-111c-4201-9f5d-ec9bf7f1d31f-kube-api-access-krpk5\") pod \"manila-operator-controller-manager-56cf9c6b99-5xzsf\" (UID: \"51327d2c-111c-4201-9f5d-ec9bf7f1d31f\") " pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.147068 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwfvm\" (UniqueName: \"kubernetes.io/projected/f4fac1bc-f504-4b82-8371-1c7afd6f76c7-kube-api-access-cwfvm\") pod \"ironic-operator-controller-manager-9fc8d5567-h5bnt\" (UID: \"f4fac1bc-f504-4b82-8371-1c7afd6f76c7\") " pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.147103 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffv29\" (UniqueName: \"kubernetes.io/projected/66342226-bc85-46a0-8d84-0627c26b196c-kube-api-access-ffv29\") pod \"horizon-operator-controller-manager-695847bc78-fcs2w\" (UID: \"66342226-bc85-46a0-8d84-0627c26b196c\") " pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.147164 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmhvw\" (UniqueName: \"kubernetes.io/projected/e5545a7b-7c61-4e6b-92e9-3a22045a2133-kube-api-access-xmhvw\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.147216 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbgl5\" (UniqueName: \"kubernetes.io/projected/aef52f91-e131-49a0-84c8-ba76db16d118-kube-api-access-pbgl5\") pod \"heat-operator-controller-manager-8ff95898-6b6gt\" (UID: \"aef52f91-e131-49a0-84c8-ba76db16d118\") " pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.147684 4727 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.147745 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert podName:e5545a7b-7c61-4e6b-92e9-3a22045a2133 nodeName:}" failed. No retries permitted until 2025-09-29 10:37:58.647724717 +0000 UTC m=+948.821038079 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert") pod "infra-operator-controller-manager-858cd69f49-swh7z" (UID: "e5545a7b-7c61-4e6b-92e9-3a22045a2133") : secret "infra-operator-webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.159949 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.168537 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.170027 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.175028 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9xk6x" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.178209 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbgl5\" (UniqueName: \"kubernetes.io/projected/aef52f91-e131-49a0-84c8-ba76db16d118-kube-api-access-pbgl5\") pod \"heat-operator-controller-manager-8ff95898-6b6gt\" (UID: \"aef52f91-e131-49a0-84c8-ba76db16d118\") " pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.198678 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.203191 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffv29\" (UniqueName: \"kubernetes.io/projected/66342226-bc85-46a0-8d84-0627c26b196c-kube-api-access-ffv29\") pod \"horizon-operator-controller-manager-695847bc78-fcs2w\" (UID: \"66342226-bc85-46a0-8d84-0627c26b196c\") " pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.203829 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.218148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmhvw\" (UniqueName: \"kubernetes.io/projected/e5545a7b-7c61-4e6b-92e9-3a22045a2133-kube-api-access-xmhvw\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.220672 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.221901 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.234897 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.235237 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-57q2p" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.235712 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.249113 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwfvm\" (UniqueName: \"kubernetes.io/projected/f4fac1bc-f504-4b82-8371-1c7afd6f76c7-kube-api-access-cwfvm\") pod \"ironic-operator-controller-manager-9fc8d5567-h5bnt\" (UID: \"f4fac1bc-f504-4b82-8371-1c7afd6f76c7\") " pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.249803 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqgzh\" (UniqueName: \"kubernetes.io/projected/ff90604b-0b99-4c43-b38a-d467d74a6376-kube-api-access-tqgzh\") pod \"neutron-operator-controller-manager-54d766c9f9-jxvz6\" (UID: \"ff90604b-0b99-4c43-b38a-d467d74a6376\") " pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.249980 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx8qk\" (UniqueName: \"kubernetes.io/projected/f0ad5209-9d60-488c-9cb7-3252bb844305-kube-api-access-bx8qk\") pod \"mariadb-operator-controller-manager-687b9cf756-z5zqv\" (UID: \"f0ad5209-9d60-488c-9cb7-3252bb844305\") " pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.250083 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4shn\" (UniqueName: \"kubernetes.io/projected/495ff530-82ab-48b5-9c58-1f03614f5b5a-kube-api-access-z4shn\") pod \"octavia-operator-controller-manager-76fcc6dc7c-47bvr\" (UID: \"495ff530-82ab-48b5-9c58-1f03614f5b5a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.250281 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g9wn\" (UniqueName: \"kubernetes.io/projected/6f3ebbd5-f00d-4db3-8160-03e9dce230e4-kube-api-access-4g9wn\") pod \"keystone-operator-controller-manager-7bf498966c-vktwk\" (UID: \"6f3ebbd5-f00d-4db3-8160-03e9dce230e4\") " pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.250428 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krpk5\" (UniqueName: \"kubernetes.io/projected/51327d2c-111c-4201-9f5d-ec9bf7f1d31f-kube-api-access-krpk5\") pod \"manila-operator-controller-manager-56cf9c6b99-5xzsf\" (UID: \"51327d2c-111c-4201-9f5d-ec9bf7f1d31f\") " pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.250531 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zv2r\" (UniqueName: \"kubernetes.io/projected/3164f156-e48b-4eaa-8e29-ecf414254d3a-kube-api-access-9zv2r\") pod \"nova-operator-controller-manager-c7c776c96-whv4l\" (UID: \"3164f156-e48b-4eaa-8e29-ecf414254d3a\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.257976 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.259203 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.261128 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-kqt4b" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.267440 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.268528 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.272720 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-brj4f" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.275584 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.289584 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krpk5\" (UniqueName: \"kubernetes.io/projected/51327d2c-111c-4201-9f5d-ec9bf7f1d31f-kube-api-access-krpk5\") pod \"manila-operator-controller-manager-56cf9c6b99-5xzsf\" (UID: \"51327d2c-111c-4201-9f5d-ec9bf7f1d31f\") " pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.289917 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.294296 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g9wn\" (UniqueName: \"kubernetes.io/projected/6f3ebbd5-f00d-4db3-8160-03e9dce230e4-kube-api-access-4g9wn\") pod \"keystone-operator-controller-manager-7bf498966c-vktwk\" (UID: \"6f3ebbd5-f00d-4db3-8160-03e9dce230e4\") " pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.294420 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwfvm\" (UniqueName: \"kubernetes.io/projected/f4fac1bc-f504-4b82-8371-1c7afd6f76c7-kube-api-access-cwfvm\") pod \"ironic-operator-controller-manager-9fc8d5567-h5bnt\" (UID: \"f4fac1bc-f504-4b82-8371-1c7afd6f76c7\") " pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.297285 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.300822 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.309951 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w96lf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.323199 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.326051 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.330064 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.330223 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.334806 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jn5lr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.353686 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx8qk\" (UniqueName: \"kubernetes.io/projected/f0ad5209-9d60-488c-9cb7-3252bb844305-kube-api-access-bx8qk\") pod \"mariadb-operator-controller-manager-687b9cf756-z5zqv\" (UID: \"f0ad5209-9d60-488c-9cb7-3252bb844305\") " pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.354067 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57kxl\" (UniqueName: \"kubernetes.io/projected/9d628165-af26-41a6-b05c-e633673213ab-kube-api-access-57kxl\") pod \"placement-operator-controller-manager-774b97b48-j7hz6\" (UID: \"9d628165-af26-41a6-b05c-e633673213ab\") " pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.354236 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4shn\" (UniqueName: \"kubernetes.io/projected/495ff530-82ab-48b5-9c58-1f03614f5b5a-kube-api-access-z4shn\") pod \"octavia-operator-controller-manager-76fcc6dc7c-47bvr\" (UID: \"495ff530-82ab-48b5-9c58-1f03614f5b5a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.354412 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.354564 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7drp\" (UniqueName: \"kubernetes.io/projected/a6dfa756-7117-4d5c-830b-a3dc23d86682-kube-api-access-b7drp\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.354894 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zv2r\" (UniqueName: \"kubernetes.io/projected/3164f156-e48b-4eaa-8e29-ecf414254d3a-kube-api-access-9zv2r\") pod \"nova-operator-controller-manager-c7c776c96-whv4l\" (UID: \"3164f156-e48b-4eaa-8e29-ecf414254d3a\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.355112 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w56fs\" (UniqueName: \"kubernetes.io/projected/19c9fecd-d2f6-453c-be12-ede6f08fcf0f-kube-api-access-w56fs\") pod \"ovn-operator-controller-manager-5f95c46c78-6rsg9\" (UID: \"19c9fecd-d2f6-453c-be12-ede6f08fcf0f\") " pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.355243 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqgzh\" (UniqueName: \"kubernetes.io/projected/ff90604b-0b99-4c43-b38a-d467d74a6376-kube-api-access-tqgzh\") pod \"neutron-operator-controller-manager-54d766c9f9-jxvz6\" (UID: \"ff90604b-0b99-4c43-b38a-d467d74a6376\") " pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.428439 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.437830 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457316 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457402 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrlww\" (UniqueName: \"kubernetes.io/projected/a64b04d3-a9a3-4c5e-a356-045f7b13c572-kube-api-access-vrlww\") pod \"telemetry-operator-controller-manager-5bf96cfbc4-zstln\" (UID: \"a64b04d3-a9a3-4c5e-a356-045f7b13c572\") " pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457475 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7drp\" (UniqueName: \"kubernetes.io/projected/a6dfa756-7117-4d5c-830b-a3dc23d86682-kube-api-access-b7drp\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457517 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w56fs\" (UniqueName: \"kubernetes.io/projected/19c9fecd-d2f6-453c-be12-ede6f08fcf0f-kube-api-access-w56fs\") pod \"ovn-operator-controller-manager-5f95c46c78-6rsg9\" (UID: \"19c9fecd-d2f6-453c-be12-ede6f08fcf0f\") " pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457618 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57kxl\" (UniqueName: \"kubernetes.io/projected/9d628165-af26-41a6-b05c-e633673213ab-kube-api-access-57kxl\") pod \"placement-operator-controller-manager-774b97b48-j7hz6\" (UID: \"9d628165-af26-41a6-b05c-e633673213ab\") " pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.457654 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s92pz\" (UniqueName: \"kubernetes.io/projected/7c7f0180-cbee-4a25-b069-3b11a90b2d08-kube-api-access-s92pz\") pod \"swift-operator-controller-manager-bc7dc7bd9-dsg68\" (UID: \"7c7f0180-cbee-4a25-b069-3b11a90b2d08\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.457904 4727 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.457959 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert podName:a6dfa756-7117-4d5c-830b-a3dc23d86682 nodeName:}" failed. No retries permitted until 2025-09-29 10:37:58.957943786 +0000 UTC m=+949.131257148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-sp6px" (UID: "a6dfa756-7117-4d5c-830b-a3dc23d86682") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.478416 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.484045 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6p49k"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.495765 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6p49k"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.495874 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.520423 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rvl8d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.526771 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7drp\" (UniqueName: \"kubernetes.io/projected/a6dfa756-7117-4d5c-830b-a3dc23d86682-kube-api-access-b7drp\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.528451 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx8qk\" (UniqueName: \"kubernetes.io/projected/f0ad5209-9d60-488c-9cb7-3252bb844305-kube-api-access-bx8qk\") pod \"mariadb-operator-controller-manager-687b9cf756-z5zqv\" (UID: \"f0ad5209-9d60-488c-9cb7-3252bb844305\") " pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.531105 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zv2r\" (UniqueName: \"kubernetes.io/projected/3164f156-e48b-4eaa-8e29-ecf414254d3a-kube-api-access-9zv2r\") pod \"nova-operator-controller-manager-c7c776c96-whv4l\" (UID: \"3164f156-e48b-4eaa-8e29-ecf414254d3a\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.532497 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4shn\" (UniqueName: \"kubernetes.io/projected/495ff530-82ab-48b5-9c58-1f03614f5b5a-kube-api-access-z4shn\") pod \"octavia-operator-controller-manager-76fcc6dc7c-47bvr\" (UID: \"495ff530-82ab-48b5-9c58-1f03614f5b5a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.536270 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.538540 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.541877 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.548163 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqgzh\" (UniqueName: \"kubernetes.io/projected/ff90604b-0b99-4c43-b38a-d467d74a6376-kube-api-access-tqgzh\") pod \"neutron-operator-controller-manager-54d766c9f9-jxvz6\" (UID: \"ff90604b-0b99-4c43-b38a-d467d74a6376\") " pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.555660 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.565411 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57kxl\" (UniqueName: \"kubernetes.io/projected/9d628165-af26-41a6-b05c-e633673213ab-kube-api-access-57kxl\") pod \"placement-operator-controller-manager-774b97b48-j7hz6\" (UID: \"9d628165-af26-41a6-b05c-e633673213ab\") " pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.565600 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gfrf8" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.566464 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s92pz\" (UniqueName: \"kubernetes.io/projected/7c7f0180-cbee-4a25-b069-3b11a90b2d08-kube-api-access-s92pz\") pod \"swift-operator-controller-manager-bc7dc7bd9-dsg68\" (UID: \"7c7f0180-cbee-4a25-b069-3b11a90b2d08\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.566505 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkhnh\" (UniqueName: \"kubernetes.io/projected/baad2063-56c9-414d-b8da-5969bf4f6220-kube-api-access-vkhnh\") pod \"test-operator-controller-manager-f66b554c6-6p49k\" (UID: \"baad2063-56c9-414d-b8da-5969bf4f6220\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.566543 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrlww\" (UniqueName: \"kubernetes.io/projected/a64b04d3-a9a3-4c5e-a356-045f7b13c572-kube-api-access-vrlww\") pod \"telemetry-operator-controller-manager-5bf96cfbc4-zstln\" (UID: \"a64b04d3-a9a3-4c5e-a356-045f7b13c572\") " pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.574226 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.574824 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w56fs\" (UniqueName: \"kubernetes.io/projected/19c9fecd-d2f6-453c-be12-ede6f08fcf0f-kube-api-access-w56fs\") pod \"ovn-operator-controller-manager-5f95c46c78-6rsg9\" (UID: \"19c9fecd-d2f6-453c-be12-ede6f08fcf0f\") " pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.597954 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.608462 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s92pz\" (UniqueName: \"kubernetes.io/projected/7c7f0180-cbee-4a25-b069-3b11a90b2d08-kube-api-access-s92pz\") pod \"swift-operator-controller-manager-bc7dc7bd9-dsg68\" (UID: \"7c7f0180-cbee-4a25-b069-3b11a90b2d08\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.622523 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.634806 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.656729 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.669115 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkhnh\" (UniqueName: \"kubernetes.io/projected/baad2063-56c9-414d-b8da-5969bf4f6220-kube-api-access-vkhnh\") pod \"test-operator-controller-manager-f66b554c6-6p49k\" (UID: \"baad2063-56c9-414d-b8da-5969bf4f6220\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.669670 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.669785 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz6hv\" (UniqueName: \"kubernetes.io/projected/fadeace3-e8ba-4e11-a2f7-6dee11f875df-kube-api-access-wz6hv\") pod \"watcher-operator-controller-manager-76669f99c-djbkf\" (UID: \"fadeace3-e8ba-4e11-a2f7-6dee11f875df\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.678970 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrlww\" (UniqueName: \"kubernetes.io/projected/a64b04d3-a9a3-4c5e-a356-045f7b13c572-kube-api-access-vrlww\") pod \"telemetry-operator-controller-manager-5bf96cfbc4-zstln\" (UID: \"a64b04d3-a9a3-4c5e-a356-045f7b13c572\") " pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.682901 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5545a7b-7c61-4e6b-92e9-3a22045a2133-cert\") pod \"infra-operator-controller-manager-858cd69f49-swh7z\" (UID: \"e5545a7b-7c61-4e6b-92e9-3a22045a2133\") " pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.689227 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.721456 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.722904 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.729276 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gxch4" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.729486 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.730794 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkhnh\" (UniqueName: \"kubernetes.io/projected/baad2063-56c9-414d-b8da-5969bf4f6220-kube-api-access-vkhnh\") pod \"test-operator-controller-manager-f66b554c6-6p49k\" (UID: \"baad2063-56c9-414d-b8da-5969bf4f6220\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.743527 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.744009 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.772094 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz6hv\" (UniqueName: \"kubernetes.io/projected/fadeace3-e8ba-4e11-a2f7-6dee11f875df-kube-api-access-wz6hv\") pod \"watcher-operator-controller-manager-76669f99c-djbkf\" (UID: \"fadeace3-e8ba-4e11-a2f7-6dee11f875df\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.772536 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.772604 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fghw4\" (UniqueName: \"kubernetes.io/projected/ca7d374b-4f00-4831-b2b1-39fbf673b820-kube-api-access-fghw4\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.788731 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.789740 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.793435 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-kjr7h" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.794201 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl"] Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.795214 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.824579 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.829645 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz6hv\" (UniqueName: \"kubernetes.io/projected/fadeace3-e8ba-4e11-a2f7-6dee11f875df-kube-api-access-wz6hv\") pod \"watcher-operator-controller-manager-76669f99c-djbkf\" (UID: \"fadeace3-e8ba-4e11-a2f7-6dee11f875df\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.849239 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.874024 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fghw4\" (UniqueName: \"kubernetes.io/projected/ca7d374b-4f00-4831-b2b1-39fbf673b820-kube-api-access-fghw4\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.874114 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkg57\" (UniqueName: \"kubernetes.io/projected/120b089a-5999-450b-acd7-5f48a5a105bd-kube-api-access-mkg57\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dgghl\" (UID: \"120b089a-5999-450b-acd7-5f48a5a105bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.874207 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.874393 4727 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: E0929 10:37:58.874452 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert podName:ca7d374b-4f00-4831-b2b1-39fbf673b820 nodeName:}" failed. No retries permitted until 2025-09-29 10:37:59.374434613 +0000 UTC m=+949.547747975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert") pod "openstack-operator-controller-manager-578979c99-jgw9d" (UID: "ca7d374b-4f00-4831-b2b1-39fbf673b820") : secret "webhook-server-cert" not found Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.882280 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.907063 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fghw4\" (UniqueName: \"kubernetes.io/projected/ca7d374b-4f00-4831-b2b1-39fbf673b820-kube-api-access-fghw4\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.982621 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkg57\" (UniqueName: \"kubernetes.io/projected/120b089a-5999-450b-acd7-5f48a5a105bd-kube-api-access-mkg57\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dgghl\" (UID: \"120b089a-5999-450b-acd7-5f48a5a105bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" Sep 29 10:37:58 crc kubenswrapper[4727]: I0929 10:37:58.982699 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.004648 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6dfa756-7117-4d5c-830b-a3dc23d86682-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-sp6px\" (UID: \"a6dfa756-7117-4d5c-830b-a3dc23d86682\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.011174 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkg57\" (UniqueName: \"kubernetes.io/projected/120b089a-5999-450b-acd7-5f48a5a105bd-kube-api-access-mkg57\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dgghl\" (UID: \"120b089a-5999-450b-acd7-5f48a5a105bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.087748 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt"] Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.198027 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.277031 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.392635 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:37:59 crc kubenswrapper[4727]: E0929 10:37:59.392923 4727 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Sep 29 10:37:59 crc kubenswrapper[4727]: E0929 10:37:59.393045 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert podName:ca7d374b-4f00-4831-b2b1-39fbf673b820 nodeName:}" failed. No retries permitted until 2025-09-29 10:38:00.393016179 +0000 UTC m=+950.566329541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert") pod "openstack-operator-controller-manager-578979c99-jgw9d" (UID: "ca7d374b-4f00-4831-b2b1-39fbf673b820") : secret "webhook-server-cert" not found Sep 29 10:37:59 crc kubenswrapper[4727]: I0929 10:37:59.607693 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.027109 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.032360 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.056985 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.087283 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.092576 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5"] Sep 29 10:38:00 crc kubenswrapper[4727]: W0929 10:38:00.107076 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c7f0180_cbee_4a25_b069_3b11a90b2d08.slice/crio-8f74efebc80e26cc543392044598b92e7c2e884314960557a0b0d492bada5a6b WatchSource:0}: Error finding container 8f74efebc80e26cc543392044598b92e7c2e884314960557a0b0d492bada5a6b: Status 404 returned error can't find the container with id 8f74efebc80e26cc543392044598b92e7c2e884314960557a0b0d492bada5a6b Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.124043 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.137501 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.194953 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.211524 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" event={"ID":"aef52f91-e131-49a0-84c8-ba76db16d118","Type":"ContainerStarted","Data":"b1a88349f1aa05e27ba0ba1c0faad1f1b4f87eee93b27efc29e6f6d8da3b318c"} Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.217428 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.222652 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.225506 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" event={"ID":"44a5b923-82b5-4af1-ad00-ba65998598ff","Type":"ContainerStarted","Data":"91bd0e3ce6debcc84b41b41e4653b1ec43e5795b84a6b2c3de3c013cefd25936"} Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.225672 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:057de94f9afa340adc34f9b25f8007d9cd2ba71bc8b5d77aac522add53b7caef,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9zv2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-c7c776c96-whv4l_openstack-operators(3164f156-e48b-4eaa-8e29-ecf414254d3a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.225787 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:bdf49c202aba5000737445bc4aeee6c5cdc6dd29c3dcd1394df9f8695830f9c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tqgzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-54d766c9f9-jxvz6_openstack-operators(ff90604b-0b99-4c43-b38a-d467d74a6376): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.225874 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z4shn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-76fcc6dc7c-47bvr_openstack-operators(495ff530-82ab-48b5-9c58-1f03614f5b5a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.226197 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vkhnh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-6p49k_openstack-operators(baad2063-56c9-414d-b8da-5969bf4f6220): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.226300 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:a7eacfe7657c55521404e56e90764896845837d62c6689b3e9485c65f99055f8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4g9wn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7bf498966c-vktwk_openstack-operators(6f3ebbd5-f00d-4db3-8160-03e9dce230e4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.226357 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b98ec0b50404626e0440bcf2e22f8d7ff06d1b1bd99f01830bceb8a2b27aa094,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwfvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-9fc8d5567-h5bnt_openstack-operators(f4fac1bc-f504-4b82-8371-1c7afd6f76c7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.228111 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z"] Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.232214 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wz6hv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-76669f99c-djbkf_openstack-operators(fadeace3-e8ba-4e11-a2f7-6dee11f875df): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.233299 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.239021 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-6p49k"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.242215 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" event={"ID":"f6f41603-6f22-4086-be87-c1a0062a691d","Type":"ContainerStarted","Data":"c1816b2cc0d1939eb6898c468fe54bee37fd542397d3112f2f56ed1e0965dc1e"} Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.246079 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.249853 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" event={"ID":"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609","Type":"ContainerStarted","Data":"fd68c7794f4a449b85033fe54256e5b9f7b53732d98bc40557494a83c82ff4f4"} Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.254466 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.255149 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" event={"ID":"9d628165-af26-41a6-b05c-e633673213ab","Type":"ContainerStarted","Data":"834b915e452dbde3503231207919080c81471f49fea8f56d23be816c1ec45914"} Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.255754 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mkg57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-dgghl_openstack-operators(120b089a-5999-450b-acd7-5f48a5a105bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.256498 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_LIGHTSPEED_IMAGE_URL_DEFAULT,Value:quay.io/openstack-lightspeed/rag-content:os-docs-2024.2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b7drp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6d776955-sp6px_openstack-operators(a6dfa756-7117-4d5c-830b-a3dc23d86682): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.256682 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf"] Sep 29 10:38:00 crc kubenswrapper[4727]: E0929 10:38:00.257069 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" podUID="120b089a-5999-450b-acd7-5f48a5a105bd" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.257245 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" event={"ID":"51327d2c-111c-4201-9f5d-ec9bf7f1d31f","Type":"ContainerStarted","Data":"4e18af753070267d9790807d273e0ad9a0e1b012ffa1f2d67156202945b649c1"} Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.259773 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" event={"ID":"19c9fecd-d2f6-453c-be12-ede6f08fcf0f","Type":"ContainerStarted","Data":"784f2a21cb71a55db1e25f7b552f41f70ffa52c44972d345942877a1ad15ad3c"} Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.262613 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.266165 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.269963 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl"] Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.376660 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w"] Sep 29 10:38:00 crc kubenswrapper[4727]: W0929 10:38:00.394208 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66342226_bc85_46a0_8d84_0627c26b196c.slice/crio-3df5604c39a41325079e81e2345379097f468c3f1bc74c06ba056faeaa9e07bd WatchSource:0}: Error finding container 3df5604c39a41325079e81e2345379097f468c3f1bc74c06ba056faeaa9e07bd: Status 404 returned error can't find the container with id 3df5604c39a41325079e81e2345379097f468c3f1bc74c06ba056faeaa9e07bd Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.420288 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.429039 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7d374b-4f00-4831-b2b1-39fbf673b820-cert\") pod \"openstack-operator-controller-manager-578979c99-jgw9d\" (UID: \"ca7d374b-4f00-4831-b2b1-39fbf673b820\") " pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.689197 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:38:00 crc kubenswrapper[4727]: I0929 10:38:00.958708 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d"] Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.288882 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" event={"ID":"a6dfa756-7117-4d5c-830b-a3dc23d86682","Type":"ContainerStarted","Data":"b3c28f1f6519a965e196f25a7d1b93d030055c281d667ca33c4e024b59f46b3d"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.293833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" event={"ID":"7c7f0180-cbee-4a25-b069-3b11a90b2d08","Type":"ContainerStarted","Data":"8f74efebc80e26cc543392044598b92e7c2e884314960557a0b0d492bada5a6b"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.295404 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" event={"ID":"3164f156-e48b-4eaa-8e29-ecf414254d3a","Type":"ContainerStarted","Data":"009a0a1401de4fd4483a4e0b41a28a7827006a27c221f0eefc110c0685de2706"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.297154 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" event={"ID":"a64b04d3-a9a3-4c5e-a356-045f7b13c572","Type":"ContainerStarted","Data":"d1d258d92ca9fd901409f4a9491d16bf10ea93a80769d51c48adee8ae254d802"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.304879 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" event={"ID":"2cb520ac-7d2f-46ab-b8bc-5f515485d278","Type":"ContainerStarted","Data":"7c47af657d9e8c5fea0084c6fcf4256a0bdf34299913f443a87d60fff02ebd03"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.308063 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" event={"ID":"66342226-bc85-46a0-8d84-0627c26b196c","Type":"ContainerStarted","Data":"3df5604c39a41325079e81e2345379097f468c3f1bc74c06ba056faeaa9e07bd"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.309453 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" event={"ID":"ff90604b-0b99-4c43-b38a-d467d74a6376","Type":"ContainerStarted","Data":"0b06095a08eafdd56cae368fcfefdd5d72780b61f3b329a298cc6caf1bcd15de"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.313609 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" event={"ID":"6f3ebbd5-f00d-4db3-8160-03e9dce230e4","Type":"ContainerStarted","Data":"f9dde4e0708f2e71fe3f06a5813678b7f84fd5d36f25db4537a59bd63f6d06db"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.319386 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" event={"ID":"baad2063-56c9-414d-b8da-5969bf4f6220","Type":"ContainerStarted","Data":"fd358cac66f7a615bab18ef2dc03654b4185f1b940eb32b0e4f87671cd4b8311"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.322034 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" event={"ID":"495ff530-82ab-48b5-9c58-1f03614f5b5a","Type":"ContainerStarted","Data":"4057d93d4758878e9c13eb7746168c68935e2253dc43a1fd78a67ed40338d3d6"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.323942 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" event={"ID":"e5545a7b-7c61-4e6b-92e9-3a22045a2133","Type":"ContainerStarted","Data":"a14257cbb893075ed20a8ab16f2692f932cb534b0297e836747617b3a4af759c"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.325487 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" event={"ID":"f4fac1bc-f504-4b82-8371-1c7afd6f76c7","Type":"ContainerStarted","Data":"c14dee6b6e85f724eda63c32d91f94ae352c57042cc224296cad64d306d1ac86"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.326993 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" event={"ID":"f0ad5209-9d60-488c-9cb7-3252bb844305","Type":"ContainerStarted","Data":"20b154a4e5290120768a41f60b72a7b122007b45a1edf37ab11a9d19dbf324b8"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.328475 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" event={"ID":"120b089a-5999-450b-acd7-5f48a5a105bd","Type":"ContainerStarted","Data":"6734435de4509520b033ad2342bb7ea50f4ab440aead68967d8eef1e5fde1395"} Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.330401 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" event={"ID":"fadeace3-e8ba-4e11-a2f7-6dee11f875df","Type":"ContainerStarted","Data":"eab4da43588bd2aae35e6b924d38857b4a9356b6ca25ba25e459fba091613631"} Sep 29 10:38:01 crc kubenswrapper[4727]: E0929 10:38:01.330647 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" podUID="120b089a-5999-450b-acd7-5f48a5a105bd" Sep 29 10:38:01 crc kubenswrapper[4727]: I0929 10:38:01.331712 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" event={"ID":"ca7d374b-4f00-4831-b2b1-39fbf673b820","Type":"ContainerStarted","Data":"82ec42c971ec6a6268e8dfe02972e4c80737e386ee2d01d47aac6f69fb91ce50"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.134399 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" podUID="3164f156-e48b-4eaa-8e29-ecf414254d3a" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.134447 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" podUID="f4fac1bc-f504-4b82-8371-1c7afd6f76c7" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.134807 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" podUID="ff90604b-0b99-4c43-b38a-d467d74a6376" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.136398 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" podUID="a6dfa756-7117-4d5c-830b-a3dc23d86682" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.138137 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" podUID="495ff530-82ab-48b5-9c58-1f03614f5b5a" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.139787 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" podUID="6f3ebbd5-f00d-4db3-8160-03e9dce230e4" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.140400 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" podUID="baad2063-56c9-414d-b8da-5969bf4f6220" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.156074 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" podUID="fadeace3-e8ba-4e11-a2f7-6dee11f875df" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.354608 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" event={"ID":"f4fac1bc-f504-4b82-8371-1c7afd6f76c7","Type":"ContainerStarted","Data":"772ccdfcd575bff34fb76e95de2610dc4b12ad96323df2f509ad044ebb6ad130"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.356305 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b98ec0b50404626e0440bcf2e22f8d7ff06d1b1bd99f01830bceb8a2b27aa094\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" podUID="f4fac1bc-f504-4b82-8371-1c7afd6f76c7" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.356982 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" event={"ID":"fadeace3-e8ba-4e11-a2f7-6dee11f875df","Type":"ContainerStarted","Data":"377f717ae37efd7cd472be4bcc7e24c0d94f3e2d20a6f9b5cf22d700823356de"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.358101 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" podUID="fadeace3-e8ba-4e11-a2f7-6dee11f875df" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.367808 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" event={"ID":"ca7d374b-4f00-4831-b2b1-39fbf673b820","Type":"ContainerStarted","Data":"153f17d32e9dd34288790b43ae74e5abd0a3316e1841a7dd5db518a9ab487ea5"} Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.372757 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" event={"ID":"6f3ebbd5-f00d-4db3-8160-03e9dce230e4","Type":"ContainerStarted","Data":"ca7ef9871b9ef45a3e1df5ba919fd28261ecaec4aab5d350662c3230cb62024d"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.374488 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:a7eacfe7657c55521404e56e90764896845837d62c6689b3e9485c65f99055f8\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" podUID="6f3ebbd5-f00d-4db3-8160-03e9dce230e4" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.415027 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" event={"ID":"3164f156-e48b-4eaa-8e29-ecf414254d3a","Type":"ContainerStarted","Data":"ee473936f4222b4b30fadb4cc28a6b48e48301827905ee814a37b9c1c29fb413"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.420598 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:057de94f9afa340adc34f9b25f8007d9cd2ba71bc8b5d77aac522add53b7caef\\\"\"" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" podUID="3164f156-e48b-4eaa-8e29-ecf414254d3a" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.431526 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" event={"ID":"495ff530-82ab-48b5-9c58-1f03614f5b5a","Type":"ContainerStarted","Data":"124810ca2ec6cb772ba1744f49fce6deb3fefd338ba32864a92bcc7b2d765c69"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.433285 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" podUID="495ff530-82ab-48b5-9c58-1f03614f5b5a" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.457509 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" event={"ID":"ff90604b-0b99-4c43-b38a-d467d74a6376","Type":"ContainerStarted","Data":"cdb2a5b3a5898096e0df441262ebb8e8143fa8f96030fbe33e958f282ed2bf46"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.460662 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:bdf49c202aba5000737445bc4aeee6c5cdc6dd29c3dcd1394df9f8695830f9c6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" podUID="ff90604b-0b99-4c43-b38a-d467d74a6376" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.465035 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" event={"ID":"a6dfa756-7117-4d5c-830b-a3dc23d86682","Type":"ContainerStarted","Data":"a2440c5eaf463466e15e87ced14fd5ae5d87c980e450941a9de641c65a67a248"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.468886 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" podUID="a6dfa756-7117-4d5c-830b-a3dc23d86682" Sep 29 10:38:02 crc kubenswrapper[4727]: I0929 10:38:02.470910 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" event={"ID":"baad2063-56c9-414d-b8da-5969bf4f6220","Type":"ContainerStarted","Data":"aba8387f87a5016a815b47995af5229893dc89156a9033d659de459697f0f4ef"} Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.473044 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" podUID="baad2063-56c9-414d-b8da-5969bf4f6220" Sep 29 10:38:02 crc kubenswrapper[4727]: E0929 10:38:02.473792 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" podUID="120b089a-5999-450b-acd7-5f48a5a105bd" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.480660 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:a7eacfe7657c55521404e56e90764896845837d62c6689b3e9485c65f99055f8\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" podUID="6f3ebbd5-f00d-4db3-8160-03e9dce230e4" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.480816 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" podUID="fadeace3-e8ba-4e11-a2f7-6dee11f875df" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.480944 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:057de94f9afa340adc34f9b25f8007d9cd2ba71bc8b5d77aac522add53b7caef\\\"\"" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" podUID="3164f156-e48b-4eaa-8e29-ecf414254d3a" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.480973 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" podUID="495ff530-82ab-48b5-9c58-1f03614f5b5a" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.481862 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" podUID="a6dfa756-7117-4d5c-830b-a3dc23d86682" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.482032 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b98ec0b50404626e0440bcf2e22f8d7ff06d1b1bd99f01830bceb8a2b27aa094\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" podUID="f4fac1bc-f504-4b82-8371-1c7afd6f76c7" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.482112 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" podUID="baad2063-56c9-414d-b8da-5969bf4f6220" Sep 29 10:38:03 crc kubenswrapper[4727]: E0929 10:38:03.483313 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:bdf49c202aba5000737445bc4aeee6c5cdc6dd29c3dcd1394df9f8695830f9c6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" podUID="ff90604b-0b99-4c43-b38a-d467d74a6376" Sep 29 10:38:04 crc kubenswrapper[4727]: I0929 10:38:04.489777 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" event={"ID":"ca7d374b-4f00-4831-b2b1-39fbf673b820","Type":"ContainerStarted","Data":"b492e17ea54b58db4e9bb59274f6584c2ebdbeab38e43c999bcf053459cef9b4"} Sep 29 10:38:04 crc kubenswrapper[4727]: I0929 10:38:04.490352 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:38:04 crc kubenswrapper[4727]: I0929 10:38:04.520479 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" podStartSLOduration=6.5204564210000004 podStartE2EDuration="6.520456421s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:38:04.515437362 +0000 UTC m=+954.688750724" watchObservedRunningTime="2025-09-29 10:38:04.520456421 +0000 UTC m=+954.693769783" Sep 29 10:38:09 crc kubenswrapper[4727]: I0929 10:38:09.523896 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" event={"ID":"aef52f91-e131-49a0-84c8-ba76db16d118","Type":"ContainerStarted","Data":"091492a627268703c2b2aef3d7ab419bfc7ef1b4bab7808be3a559cc6fbb4b6f"} Sep 29 10:38:10 crc kubenswrapper[4727]: I0929 10:38:10.695742 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-578979c99-jgw9d" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.553822 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" event={"ID":"66342226-bc85-46a0-8d84-0627c26b196c","Type":"ContainerStarted","Data":"99e0308ebb7964aa502b10e7fd4737f92f681d640119f7d605ac1f54e01c95ef"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.554162 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" event={"ID":"66342226-bc85-46a0-8d84-0627c26b196c","Type":"ContainerStarted","Data":"2e95357fe0bcc2eb5045f1b0688019de132073c4246c05b5314227fbf5da828f"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.555123 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.556164 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" event={"ID":"19c9fecd-d2f6-453c-be12-ede6f08fcf0f","Type":"ContainerStarted","Data":"be7cfccda293e1e418a2a3e23acea79c6b74057101f8cef090909bddec1e6f15"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.557133 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" event={"ID":"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609","Type":"ContainerStarted","Data":"ce9092bcc6962c0409de5b2ee21c3469765e1b060375b90a14b3b8ee4f822d0b"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.558052 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" event={"ID":"9d628165-af26-41a6-b05c-e633673213ab","Type":"ContainerStarted","Data":"9b5e82a5bbce525162ce17963289f2cf0239caef86945468485e3e8d114ea46f"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.558939 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" event={"ID":"f6f41603-6f22-4086-be87-c1a0062a691d","Type":"ContainerStarted","Data":"ccfd2950b9d395ec57d8d56dd10de0cd13c27a20043523cd962088c48dfc2ba9"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.559817 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" event={"ID":"2cb520ac-7d2f-46ab-b8bc-5f515485d278","Type":"ContainerStarted","Data":"13734d74b8920e12f16fe3841f70c15b04ee5d97fbdb5798d7b87d82e8a0710f"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.567932 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" event={"ID":"7c7f0180-cbee-4a25-b069-3b11a90b2d08","Type":"ContainerStarted","Data":"13aae7b65fb228244a3d5e158b99052a96b972d2779b5568eb27447a94161d6b"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.596803 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" event={"ID":"51327d2c-111c-4201-9f5d-ec9bf7f1d31f","Type":"ContainerStarted","Data":"850b2e2b06d1f7ffb2d098947d0069f3444afd120ab3bfb7aaf2c0dced375f09"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.598569 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" event={"ID":"aef52f91-e131-49a0-84c8-ba76db16d118","Type":"ContainerStarted","Data":"1bc654a816000b194cd77a2680d25233827e625f3feb630d0747dda127749c97"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.603936 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.606621 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" podStartSLOduration=4.330803515 podStartE2EDuration="16.606588144s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.403682838 +0000 UTC m=+950.576996200" lastFinishedPulling="2025-09-29 10:38:12.679467467 +0000 UTC m=+962.852780829" observedRunningTime="2025-09-29 10:38:13.604257644 +0000 UTC m=+963.777571006" watchObservedRunningTime="2025-09-29 10:38:13.606588144 +0000 UTC m=+963.779901516" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.608033 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" event={"ID":"44a5b923-82b5-4af1-ad00-ba65998598ff","Type":"ContainerStarted","Data":"2d635eaae749cd51ebdd5d31b875cdfc9b792429d35ecfefd162492dd0ddc759"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.608650 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.633655 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" event={"ID":"a64b04d3-a9a3-4c5e-a356-045f7b13c572","Type":"ContainerStarted","Data":"4fc62ceca495bdd9a0b413247ba48bbec3a9eb327ba6a7e309cf22a2ceb2d201"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.647075 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8ff95898-6b6gt" podStartSLOduration=12.174316675 podStartE2EDuration="16.64448861s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:37:59.20359929 +0000 UTC m=+949.376912652" lastFinishedPulling="2025-09-29 10:38:03.673771225 +0000 UTC m=+953.847084587" observedRunningTime="2025-09-29 10:38:13.641153975 +0000 UTC m=+963.814467337" watchObservedRunningTime="2025-09-29 10:38:13.64448861 +0000 UTC m=+963.817801972" Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.667422 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" event={"ID":"f0ad5209-9d60-488c-9cb7-3252bb844305","Type":"ContainerStarted","Data":"d374587a4a2c033eb4dd4431502f939f1828b017c7025d9245d25f6c53592cbb"} Sep 29 10:38:13 crc kubenswrapper[4727]: I0929 10:38:13.697353 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" event={"ID":"e5545a7b-7c61-4e6b-92e9-3a22045a2133","Type":"ContainerStarted","Data":"e354777437f0f61862708a5694e8b7b8814dc14a488b35c5b03c35ec65fc4b4d"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.729624 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" event={"ID":"19c9fecd-d2f6-453c-be12-ede6f08fcf0f","Type":"ContainerStarted","Data":"b6d5c031a444db09a36f950546a78c5468281501aa7ea0c0b8b32af607d66ce4"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.730137 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.731727 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" event={"ID":"2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609","Type":"ContainerStarted","Data":"1d6f3e67d04d8fe1dfe8a56c14f50d3bf0c57d3f4ab2fad0b5f6dddd5d84d0f9"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.731837 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.733370 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" event={"ID":"51327d2c-111c-4201-9f5d-ec9bf7f1d31f","Type":"ContainerStarted","Data":"26612a1504d79da292f2be3122b5cda46845c2f15e0b0315d3ac27b41e7e1a24"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.733519 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.738948 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" event={"ID":"44a5b923-82b5-4af1-ad00-ba65998598ff","Type":"ContainerStarted","Data":"35bed1b92f9cee44297e3263ae559de439135a342bac2c3a3da1ed18b0ce5a88"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.739120 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.741412 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" event={"ID":"2cb520ac-7d2f-46ab-b8bc-5f515485d278","Type":"ContainerStarted","Data":"a508a248afa7fa9e540efb8d467361f8b49336f94af096dd83d81c13b2ad6c72"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.741632 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.743740 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" event={"ID":"f0ad5209-9d60-488c-9cb7-3252bb844305","Type":"ContainerStarted","Data":"4d83f485814ab07e17a4d5dd56a5faff908dd3c3807e8ab4efdbd76f5b13dc75"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.743848 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.745888 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" event={"ID":"7c7f0180-cbee-4a25-b069-3b11a90b2d08","Type":"ContainerStarted","Data":"e8f1bbe6df35a65cad17e9ce76906afed682ca6af74a401eefaef01ed3fd19ee"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.746030 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.748426 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" event={"ID":"f6f41603-6f22-4086-be87-c1a0062a691d","Type":"ContainerStarted","Data":"744a85e640518587e0431e1439cff6565e61c23f1b65d68b6ef640f79159bd73"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.749100 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.751224 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" event={"ID":"e5545a7b-7c61-4e6b-92e9-3a22045a2133","Type":"ContainerStarted","Data":"5b40434b4f2833a96b9f327c3513235dcda35d93cd637c8f035dd5b260bcd813"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.751314 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.753665 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" event={"ID":"9d628165-af26-41a6-b05c-e633673213ab","Type":"ContainerStarted","Data":"22b548f06ce787cd81327d0f1e4c4d8c13be29f96420608cb990d3086d99a821"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.753843 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.756017 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" event={"ID":"a64b04d3-a9a3-4c5e-a356-045f7b13c572","Type":"ContainerStarted","Data":"d34b0b08e2a2aa1c135117ce1f682da6becdf5e44846331e9b13b783a4a5b7b0"} Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.756377 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.789957 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" podStartSLOduration=4.164103075 podStartE2EDuration="16.78992799s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.066719159 +0000 UTC m=+950.240032521" lastFinishedPulling="2025-09-29 10:38:12.692544064 +0000 UTC m=+962.865857436" observedRunningTime="2025-09-29 10:38:14.762800242 +0000 UTC m=+964.936113604" watchObservedRunningTime="2025-09-29 10:38:14.78992799 +0000 UTC m=+964.963241352" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.790466 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" podStartSLOduration=5.29243968 podStartE2EDuration="17.790460034s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.169777343 +0000 UTC m=+950.343090705" lastFinishedPulling="2025-09-29 10:38:12.667797697 +0000 UTC m=+962.841111059" observedRunningTime="2025-09-29 10:38:14.788580316 +0000 UTC m=+964.961893678" watchObservedRunningTime="2025-09-29 10:38:14.790460034 +0000 UTC m=+964.963773396" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.820886 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" podStartSLOduration=4.7566007809999995 podStartE2EDuration="17.820859547s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:37:59.618209218 +0000 UTC m=+949.791522580" lastFinishedPulling="2025-09-29 10:38:12.682467984 +0000 UTC m=+962.855781346" observedRunningTime="2025-09-29 10:38:14.816925526 +0000 UTC m=+964.990238888" watchObservedRunningTime="2025-09-29 10:38:14.820859547 +0000 UTC m=+964.994172909" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.847176 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" podStartSLOduration=5.353526634 podStartE2EDuration="17.847147664s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.183189069 +0000 UTC m=+950.356502431" lastFinishedPulling="2025-09-29 10:38:12.676810109 +0000 UTC m=+962.850123461" observedRunningTime="2025-09-29 10:38:14.840606196 +0000 UTC m=+965.013919558" watchObservedRunningTime="2025-09-29 10:38:14.847147664 +0000 UTC m=+965.020461026" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.867220 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" podStartSLOduration=5.390572588 podStartE2EDuration="17.86719499s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.191217506 +0000 UTC m=+950.364530868" lastFinishedPulling="2025-09-29 10:38:12.667839908 +0000 UTC m=+962.841153270" observedRunningTime="2025-09-29 10:38:14.863107765 +0000 UTC m=+965.036421127" watchObservedRunningTime="2025-09-29 10:38:14.86719499 +0000 UTC m=+965.040508352" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.881043 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" podStartSLOduration=5.298130688 podStartE2EDuration="17.881000406s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.0939202 +0000 UTC m=+950.267233562" lastFinishedPulling="2025-09-29 10:38:12.676789918 +0000 UTC m=+962.850103280" observedRunningTime="2025-09-29 10:38:14.878259895 +0000 UTC m=+965.051573257" watchObservedRunningTime="2025-09-29 10:38:14.881000406 +0000 UTC m=+965.054313798" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.910798 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" podStartSLOduration=4.412035101 podStartE2EDuration="16.910771333s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.139414551 +0000 UTC m=+950.312727913" lastFinishedPulling="2025-09-29 10:38:12.638150783 +0000 UTC m=+962.811464145" observedRunningTime="2025-09-29 10:38:14.906424661 +0000 UTC m=+965.079738033" watchObservedRunningTime="2025-09-29 10:38:14.910771333 +0000 UTC m=+965.084084695" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.938028 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" podStartSLOduration=5.380632022 podStartE2EDuration="17.937974723s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.080963286 +0000 UTC m=+950.254276648" lastFinishedPulling="2025-09-29 10:38:12.638305987 +0000 UTC m=+962.811619349" observedRunningTime="2025-09-29 10:38:14.937655985 +0000 UTC m=+965.110969347" watchObservedRunningTime="2025-09-29 10:38:14.937974723 +0000 UTC m=+965.111288085" Sep 29 10:38:14 crc kubenswrapper[4727]: I0929 10:38:14.997939 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" podStartSLOduration=4.393133894 podStartE2EDuration="16.997905777s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.072079037 +0000 UTC m=+950.245392399" lastFinishedPulling="2025-09-29 10:38:12.67685092 +0000 UTC m=+962.850164282" observedRunningTime="2025-09-29 10:38:14.964436665 +0000 UTC m=+965.137750037" watchObservedRunningTime="2025-09-29 10:38:14.997905777 +0000 UTC m=+965.171219139" Sep 29 10:38:15 crc kubenswrapper[4727]: I0929 10:38:15.002021 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" podStartSLOduration=4.521556032 podStartE2EDuration="17.002011023s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.187345386 +0000 UTC m=+950.360658748" lastFinishedPulling="2025-09-29 10:38:12.667800377 +0000 UTC m=+962.841113739" observedRunningTime="2025-09-29 10:38:14.985006355 +0000 UTC m=+965.158319717" watchObservedRunningTime="2025-09-29 10:38:15.002011023 +0000 UTC m=+965.175324385" Sep 29 10:38:15 crc kubenswrapper[4727]: I0929 10:38:15.011947 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" podStartSLOduration=5.411101947 podStartE2EDuration="18.011927488s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.067078048 +0000 UTC m=+950.240391410" lastFinishedPulling="2025-09-29 10:38:12.667903589 +0000 UTC m=+962.841216951" observedRunningTime="2025-09-29 10:38:15.011575689 +0000 UTC m=+965.184889051" watchObservedRunningTime="2025-09-29 10:38:15.011927488 +0000 UTC m=+965.185240850" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.100049 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6495d75b5-rbtdb" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.101716 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-748c574d75-8s5z5" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.130756 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d74f4d695-8txnk" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.139893 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-67b5d44b7f-vx2tw" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.485084 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-695847bc78-fcs2w" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.546158 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-56cf9c6b99-5xzsf" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.562351 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-687b9cf756-z5zqv" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.663392 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-dsg68" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.700839 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5f95c46c78-6rsg9" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.756646 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-774b97b48-j7hz6" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.813740 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5bf96cfbc4-zstln" Sep 29 10:38:18 crc kubenswrapper[4727]: I0929 10:38:18.894528 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-858cd69f49-swh7z" Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.246713 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.246775 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.246827 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.247499 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.247581 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435" gracePeriod=600 Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.800198 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435" exitCode=0 Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.800237 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435"} Sep 29 10:38:19 crc kubenswrapper[4727]: I0929 10:38:19.800267 4727 scope.go:117] "RemoveContainer" containerID="2b9861fc098d24823ee0e061cc929b4101985e872b5b831e196a86ad3377e090" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.853686 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" event={"ID":"6f3ebbd5-f00d-4db3-8160-03e9dce230e4","Type":"ContainerStarted","Data":"686f421ecdb8abae03c0053db44c65145393bafada8e7a108d4d21371e99425e"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.854413 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.856422 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" event={"ID":"baad2063-56c9-414d-b8da-5969bf4f6220","Type":"ContainerStarted","Data":"1f9b7ffef22c08b55d069821447ed023544c3e6ebf659f6c9be45d4ac51ab4e6"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.856564 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.858024 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" event={"ID":"3164f156-e48b-4eaa-8e29-ecf414254d3a","Type":"ContainerStarted","Data":"881d00ab1fba45f5bdb88580c19c73999c8b644ff089a684ae732dd228685e46"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.858188 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.860168 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" event={"ID":"495ff530-82ab-48b5-9c58-1f03614f5b5a","Type":"ContainerStarted","Data":"0b473960e6e0b4df6e6cd1d1fe1cd4b69559c769a51da2bbdd27285dfef39a68"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.860374 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.863399 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" event={"ID":"f4fac1bc-f504-4b82-8371-1c7afd6f76c7","Type":"ContainerStarted","Data":"9fa0cab902c4a9c0e7251e4873d4d087c30fa164486b790ea61affc60990170c"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.863616 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.865350 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" event={"ID":"fadeace3-e8ba-4e11-a2f7-6dee11f875df","Type":"ContainerStarted","Data":"db7f454f047e34c59a32764b8f7b03d1d64db9b76e585581d85492e0775c3866"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.865752 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.868640 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" event={"ID":"120b089a-5999-450b-acd7-5f48a5a105bd","Type":"ContainerStarted","Data":"4f7559a3d61751adaba378b953e6e8d0cf3265022decb1331f2f8e324bb9c5af"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.870490 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" podStartSLOduration=6.30745633 podStartE2EDuration="28.870473084s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.226238158 +0000 UTC m=+950.399551520" lastFinishedPulling="2025-09-29 10:38:22.789254912 +0000 UTC m=+972.962568274" observedRunningTime="2025-09-29 10:38:25.868900974 +0000 UTC m=+976.042214356" watchObservedRunningTime="2025-09-29 10:38:25.870473084 +0000 UTC m=+976.043786446" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.875989 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" event={"ID":"a6dfa756-7117-4d5c-830b-a3dc23d86682","Type":"ContainerStarted","Data":"5d9433a337e18d7793f01ce69be6839a45aaa40182391decb3024d9965b046f2"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.876392 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.880502 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.883322 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" podStartSLOduration=3.3649924909999998 podStartE2EDuration="27.883305333s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.225821967 +0000 UTC m=+950.399135329" lastFinishedPulling="2025-09-29 10:38:24.744134809 +0000 UTC m=+974.917448171" observedRunningTime="2025-09-29 10:38:25.88280234 +0000 UTC m=+976.056115712" watchObservedRunningTime="2025-09-29 10:38:25.883305333 +0000 UTC m=+976.056618695" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.884648 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" event={"ID":"ff90604b-0b99-4c43-b38a-d467d74a6376","Type":"ContainerStarted","Data":"efde812a3087406198482f8928e102942c509cf55e85e156e9b4ce18a43dd77e"} Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.885061 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.923066 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" podStartSLOduration=4.8556792909999995 podStartE2EDuration="28.923043002s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.226243768 +0000 UTC m=+950.399557130" lastFinishedPulling="2025-09-29 10:38:24.293607479 +0000 UTC m=+974.466920841" observedRunningTime="2025-09-29 10:38:25.921597185 +0000 UTC m=+976.094910557" watchObservedRunningTime="2025-09-29 10:38:25.923043002 +0000 UTC m=+976.096356364" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.927388 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" podStartSLOduration=3.439218775 podStartE2EDuration="27.927367573s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.226125335 +0000 UTC m=+950.399438697" lastFinishedPulling="2025-09-29 10:38:24.714274133 +0000 UTC m=+974.887587495" observedRunningTime="2025-09-29 10:38:25.902317741 +0000 UTC m=+976.075631103" watchObservedRunningTime="2025-09-29 10:38:25.927367573 +0000 UTC m=+976.100680945" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.937200 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dgghl" podStartSLOduration=3.377831613 podStartE2EDuration="27.937175894s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.255538902 +0000 UTC m=+950.428852264" lastFinishedPulling="2025-09-29 10:38:24.814883183 +0000 UTC m=+974.988196545" observedRunningTime="2025-09-29 10:38:25.935077091 +0000 UTC m=+976.108390453" watchObservedRunningTime="2025-09-29 10:38:25.937175894 +0000 UTC m=+976.110489286" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.956768 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" podStartSLOduration=6.393111535 podStartE2EDuration="28.956746416s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.225527709 +0000 UTC m=+950.398841071" lastFinishedPulling="2025-09-29 10:38:22.78916259 +0000 UTC m=+972.962475952" observedRunningTime="2025-09-29 10:38:25.953520853 +0000 UTC m=+976.126834215" watchObservedRunningTime="2025-09-29 10:38:25.956746416 +0000 UTC m=+976.130059778" Sep 29 10:38:25 crc kubenswrapper[4727]: I0929 10:38:25.972570 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" podStartSLOduration=5.414691509 podStartE2EDuration="27.972555321s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.232081468 +0000 UTC m=+950.405394830" lastFinishedPulling="2025-09-29 10:38:22.78994528 +0000 UTC m=+972.963258642" observedRunningTime="2025-09-29 10:38:25.9709577 +0000 UTC m=+976.144271062" watchObservedRunningTime="2025-09-29 10:38:25.972555321 +0000 UTC m=+976.145868683" Sep 29 10:38:26 crc kubenswrapper[4727]: I0929 10:38:26.016490 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" podStartSLOduration=3.561506333 podStartE2EDuration="28.016468597s" podCreationTimestamp="2025-09-29 10:37:58 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.255913042 +0000 UTC m=+950.429226404" lastFinishedPulling="2025-09-29 10:38:24.710875306 +0000 UTC m=+974.884188668" observedRunningTime="2025-09-29 10:38:26.013612594 +0000 UTC m=+976.186925976" watchObservedRunningTime="2025-09-29 10:38:26.016468597 +0000 UTC m=+976.189781959" Sep 29 10:38:26 crc kubenswrapper[4727]: I0929 10:38:26.018760 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" podStartSLOduration=6.45529966 podStartE2EDuration="29.018750706s" podCreationTimestamp="2025-09-29 10:37:57 +0000 UTC" firstStartedPulling="2025-09-29 10:38:00.225725644 +0000 UTC m=+950.399039006" lastFinishedPulling="2025-09-29 10:38:22.78917669 +0000 UTC m=+972.962490052" observedRunningTime="2025-09-29 10:38:25.986109899 +0000 UTC m=+976.159423261" watchObservedRunningTime="2025-09-29 10:38:26.018750706 +0000 UTC m=+976.192064068" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.330725 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9fc8d5567-h5bnt" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.441506 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7bf498966c-vktwk" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.577829 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54d766c9f9-jxvz6" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.603451 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-whv4l" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.638944 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-47bvr" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.828436 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-6p49k" Sep 29 10:38:38 crc kubenswrapper[4727]: I0929 10:38:38.853755 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-djbkf" Sep 29 10:38:39 crc kubenswrapper[4727]: I0929 10:38:39.284371 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-sp6px" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.181268 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.183016 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.185654 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.185652 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.185839 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-vk84r" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.186231 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.196767 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.263761 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.265909 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58jkg\" (UniqueName: \"kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.265986 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.281573 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.284620 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.295725 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.367560 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2xx\" (UniqueName: \"kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.367629 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58jkg\" (UniqueName: \"kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.367802 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.367918 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.368166 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.368881 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.391084 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58jkg\" (UniqueName: \"kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg\") pod \"dnsmasq-dns-675f4bcbfc-rbhdz\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.469660 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.469758 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.469797 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2xx\" (UniqueName: \"kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.470767 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.470846 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.488993 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2xx\" (UniqueName: \"kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx\") pod \"dnsmasq-dns-78dd6ddcc-n98xj\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.505843 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.608643 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.848558 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.848768 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:38:56 crc kubenswrapper[4727]: I0929 10:38:56.924734 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:38:56 crc kubenswrapper[4727]: W0929 10:38:56.929960 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe36e2ed_2475_4b53_8d00_ae2b616cfac0.slice/crio-e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae WatchSource:0}: Error finding container e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae: Status 404 returned error can't find the container with id e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae Sep 29 10:38:57 crc kubenswrapper[4727]: I0929 10:38:57.118352 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" event={"ID":"dd80e64b-eafd-40c9-8f06-6233708a7ee5","Type":"ContainerStarted","Data":"8272a424752ddc856a90576b23566cdca9b1b83a08ee7a4237511cd69e19bc2a"} Sep 29 10:38:57 crc kubenswrapper[4727]: I0929 10:38:57.119913 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" event={"ID":"be36e2ed-2475-4b53-8d00-ae2b616cfac0","Type":"ContainerStarted","Data":"e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae"} Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.296170 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.343417 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.352509 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.360891 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.521157 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.521729 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.521756 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b75fl\" (UniqueName: \"kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.622943 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.623048 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.623077 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b75fl\" (UniqueName: \"kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.623913 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.624278 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.664460 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b75fl\" (UniqueName: \"kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl\") pod \"dnsmasq-dns-5ccc8479f9-w6pdj\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.688407 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.713473 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.748538 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.752492 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.765954 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.825825 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.825877 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.826109 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-622ws\" (UniqueName: \"kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.929070 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-622ws\" (UniqueName: \"kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.929135 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.929161 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.930107 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.930703 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:38:59 crc kubenswrapper[4727]: I0929 10:38:59.966289 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-622ws\" (UniqueName: \"kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws\") pod \"dnsmasq-dns-57d769cc4f-524xh\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.163098 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.281738 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.567696 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.569059 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.571130 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.571468 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.573117 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-48nph" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.573497 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.573651 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.573834 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.573982 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.600277 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.743973 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh8xp\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744324 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744389 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744410 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744444 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744497 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744515 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744535 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744568 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744588 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.744610 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.846311 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.846937 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.846975 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847016 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh8xp\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847053 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847119 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847141 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847171 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847200 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847223 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847248 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.846804 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.847978 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.848612 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.849093 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.849148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.849285 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.852260 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.861247 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.866193 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.869750 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.878551 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.881043 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.881990 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.882645 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.883321 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.883528 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.884657 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.884862 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.884991 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.885062 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.885402 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8kslj" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.888911 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh8xp\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:00 crc kubenswrapper[4727]: I0929 10:39:00.899132 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052174 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052317 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052419 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052457 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pl5l\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052537 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052665 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052703 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052738 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052771 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052807 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.052840 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154743 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154793 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154854 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154889 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154918 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.154985 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155028 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pl5l\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155063 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155062 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155091 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155109 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.155130 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.156282 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.157088 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.157746 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.158155 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.158983 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.161862 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.162553 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.167223 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.170723 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.173687 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pl5l\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.182950 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " pod="openstack/rabbitmq-server-0" Sep 29 10:39:01 crc kubenswrapper[4727]: I0929 10:39:01.264290 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.469142 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.470825 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.475981 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.476400 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.477109 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.477163 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6n2mp" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.477280 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.501819 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.512014 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609285 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-secrets\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609434 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609492 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609512 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609701 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wkl9\" (UniqueName: \"kubernetes.io/projected/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kube-api-access-9wkl9\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609774 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.609966 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.610039 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.610181 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711672 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711807 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-secrets\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711825 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711852 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711878 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711918 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wkl9\" (UniqueName: \"kubernetes.io/projected/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kube-api-access-9wkl9\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711940 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.711985 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.712635 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.713078 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.713308 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kolla-config\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.713580 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-config-data-default\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.714379 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.717588 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-secrets\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.718690 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.719023 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.734309 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.736020 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wkl9\" (UniqueName: \"kubernetes.io/projected/2d25415e-3c8d-4cd3-ab20-83a815a5e39f-kube-api-access-9wkl9\") pod \"openstack-galera-0\" (UID: \"2d25415e-3c8d-4cd3-ab20-83a815a5e39f\") " pod="openstack/openstack-galera-0" Sep 29 10:39:02 crc kubenswrapper[4727]: I0929 10:39:02.841670 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.576972 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.578812 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.581065 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8cgkw" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.581127 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.584654 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.601306 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.606728 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731132 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731430 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731460 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731493 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mf7l\" (UniqueName: \"kubernetes.io/projected/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kube-api-access-8mf7l\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731540 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731644 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.731697 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.732042 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.732116 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833233 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833296 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833355 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833395 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833418 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833444 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mf7l\" (UniqueName: \"kubernetes.io/projected/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kube-api-access-8mf7l\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833466 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833509 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833594 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.833967 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.837125 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.838521 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.839239 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.839728 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.845617 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.851826 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.853514 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9728f6-fdb6-47ba-a17c-4220301c2a88-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.871216 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.871778 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mf7l\" (UniqueName: \"kubernetes.io/projected/ac9728f6-fdb6-47ba-a17c-4220301c2a88-kube-api-access-8mf7l\") pod \"openstack-cell1-galera-0\" (UID: \"ac9728f6-fdb6-47ba-a17c-4220301c2a88\") " pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.907755 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.909645 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.912802 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ptwhb" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.913213 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.914304 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.923754 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:03 crc kubenswrapper[4727]: I0929 10:39:03.931574 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.036069 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-config-data\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.036131 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kolla-config\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.036187 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.036263 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.036357 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69vdd\" (UniqueName: \"kubernetes.io/projected/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kube-api-access-69vdd\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.137769 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.137855 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69vdd\" (UniqueName: \"kubernetes.io/projected/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kube-api-access-69vdd\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.137899 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-config-data\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.137921 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kolla-config\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.137949 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.138745 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-config-data\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.138815 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kolla-config\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.143208 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.144138 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.157602 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69vdd\" (UniqueName: \"kubernetes.io/projected/8f2d7f61-fe27-4af2-8ef4-05d247fde09d-kube-api-access-69vdd\") pod \"memcached-0\" (UID: \"8f2d7f61-fe27-4af2-8ef4-05d247fde09d\") " pod="openstack/memcached-0" Sep 29 10:39:04 crc kubenswrapper[4727]: I0929 10:39:04.254534 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 29 10:39:05 crc kubenswrapper[4727]: I0929 10:39:05.825661 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:39:05 crc kubenswrapper[4727]: I0929 10:39:05.827466 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:39:05 crc kubenswrapper[4727]: I0929 10:39:05.829642 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5phdn" Sep 29 10:39:05 crc kubenswrapper[4727]: I0929 10:39:05.837037 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:39:05 crc kubenswrapper[4727]: I0929 10:39:05.967504 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zp8f\" (UniqueName: \"kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f\") pod \"kube-state-metrics-0\" (UID: \"03aebbc4-3f74-46f5-84b9-c0b7935569b7\") " pod="openstack/kube-state-metrics-0" Sep 29 10:39:06 crc kubenswrapper[4727]: I0929 10:39:06.069490 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zp8f\" (UniqueName: \"kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f\") pod \"kube-state-metrics-0\" (UID: \"03aebbc4-3f74-46f5-84b9-c0b7935569b7\") " pod="openstack/kube-state-metrics-0" Sep 29 10:39:06 crc kubenswrapper[4727]: I0929 10:39:06.097107 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zp8f\" (UniqueName: \"kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f\") pod \"kube-state-metrics-0\" (UID: \"03aebbc4-3f74-46f5-84b9-c0b7935569b7\") " pod="openstack/kube-state-metrics-0" Sep 29 10:39:06 crc kubenswrapper[4727]: I0929 10:39:06.144594 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:39:07 crc kubenswrapper[4727]: I0929 10:39:07.208057 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" event={"ID":"bd664253-6a4b-4b3f-baba-ad5ef08b8485","Type":"ContainerStarted","Data":"05d92e582a33338b3ad469960381f09b12f79e8fd14ae3054919c1978b0fd141"} Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.196380 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.199463 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.207101 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.208146 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.210297 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.210646 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.210778 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.210833 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-9pcgf" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.317714 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.317795 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.317979 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.318058 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.318080 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-config\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.318140 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzzj\" (UniqueName: \"kubernetes.io/projected/4bb550f2-83cb-472e-9e1e-b5da4779b42d-kube-api-access-2qzzj\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.318741 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.318793 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440510 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440573 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440633 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440672 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-config\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440768 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.440840 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzzj\" (UniqueName: \"kubernetes.io/projected/4bb550f2-83cb-472e-9e1e-b5da4779b42d-kube-api-access-2qzzj\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.442041 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.442252 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.443068 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.445937 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb550f2-83cb-472e-9e1e-b5da4779b42d-config\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.458505 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.460221 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.462900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzzj\" (UniqueName: \"kubernetes.io/projected/4bb550f2-83cb-472e-9e1e-b5da4779b42d-kube-api-access-2qzzj\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.464452 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb550f2-83cb-472e-9e1e-b5da4779b42d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.469811 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4bb550f2-83cb-472e-9e1e-b5da4779b42d\") " pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:09 crc kubenswrapper[4727]: I0929 10:39:09.527032 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.246999 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x86c4"] Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.248288 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.251035 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-v4k89" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.251130 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.251173 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.271382 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fhhrr"] Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.273043 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.288368 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4"] Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.335449 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fhhrr"] Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353252 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-run\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353645 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq2nt\" (UniqueName: \"kubernetes.io/projected/36e70287-3b92-41e6-a056-fb29d1e03772-kube-api-access-xq2nt\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353686 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36e70287-3b92-41e6-a056-fb29d1e03772-scripts\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353714 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353747 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-log\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353832 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-lib\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.353905 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-log-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354045 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-etc-ovs\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354165 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354193 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-ovn-controller-tls-certs\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354259 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkghf\" (UniqueName: \"kubernetes.io/projected/7cb18806-a80a-491a-8ade-6371af8b54a9-kube-api-access-wkghf\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354298 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-combined-ca-bundle\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.354386 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cb18806-a80a-491a-8ade-6371af8b54a9-scripts\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.455940 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.455994 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-ovn-controller-tls-certs\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456047 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkghf\" (UniqueName: \"kubernetes.io/projected/7cb18806-a80a-491a-8ade-6371af8b54a9-kube-api-access-wkghf\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456080 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-combined-ca-bundle\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456125 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cb18806-a80a-491a-8ade-6371af8b54a9-scripts\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456179 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-run\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456213 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq2nt\" (UniqueName: \"kubernetes.io/projected/36e70287-3b92-41e6-a056-fb29d1e03772-kube-api-access-xq2nt\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456250 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36e70287-3b92-41e6-a056-fb29d1e03772-scripts\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456272 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456303 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-log\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456363 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-lib\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456395 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-log-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456422 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-etc-ovs\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456788 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-run\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456797 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-etc-ovs\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456857 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456923 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-run-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456958 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36e70287-3b92-41e6-a056-fb29d1e03772-var-log-ovn\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.456988 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-log\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.457616 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7cb18806-a80a-491a-8ade-6371af8b54a9-var-lib\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.458445 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cb18806-a80a-491a-8ade-6371af8b54a9-scripts\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.459289 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36e70287-3b92-41e6-a056-fb29d1e03772-scripts\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.460252 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-combined-ca-bundle\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.473667 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e70287-3b92-41e6-a056-fb29d1e03772-ovn-controller-tls-certs\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.482815 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq2nt\" (UniqueName: \"kubernetes.io/projected/36e70287-3b92-41e6-a056-fb29d1e03772-kube-api-access-xq2nt\") pod \"ovn-controller-x86c4\" (UID: \"36e70287-3b92-41e6-a056-fb29d1e03772\") " pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.484042 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkghf\" (UniqueName: \"kubernetes.io/projected/7cb18806-a80a-491a-8ade-6371af8b54a9-kube-api-access-wkghf\") pod \"ovn-controller-ovs-fhhrr\" (UID: \"7cb18806-a80a-491a-8ade-6371af8b54a9\") " pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.581060 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4" Sep 29 10:39:10 crc kubenswrapper[4727]: I0929 10:39:10.593707 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:11 crc kubenswrapper[4727]: I0929 10:39:11.400010 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:39:11 crc kubenswrapper[4727]: I0929 10:39:11.448470 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:39:11 crc kubenswrapper[4727]: W0929 10:39:11.909515 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod667cfa10_bc94_4788_af5d_296745385383.slice/crio-b910c6dbfec64243cf3bf137a89f4a36681499fae1b69cafb5f2697e92dfcdee WatchSource:0}: Error finding container b910c6dbfec64243cf3bf137a89f4a36681499fae1b69cafb5f2697e92dfcdee: Status 404 returned error can't find the container with id b910c6dbfec64243cf3bf137a89f4a36681499fae1b69cafb5f2697e92dfcdee Sep 29 10:39:11 crc kubenswrapper[4727]: W0929 10:39:11.911720 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03aebbc4_3f74_46f5_84b9_c0b7935569b7.slice/crio-9684e84219ccd0abdc76e486a7541f648e284063a4970f063b2f930748080083 WatchSource:0}: Error finding container 9684e84219ccd0abdc76e486a7541f648e284063a4970f063b2f930748080083: Status 404 returned error can't find the container with id 9684e84219ccd0abdc76e486a7541f648e284063a4970f063b2f930748080083 Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.924558 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.924722 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58jkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-rbhdz_openstack(be36e2ed-2475-4b53-8d00-ae2b616cfac0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.926362 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" podUID="be36e2ed-2475-4b53-8d00-ae2b616cfac0" Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.932163 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.932420 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jl2xx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-n98xj_openstack(dd80e64b-eafd-40c9-8f06-6233708a7ee5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:39:11 crc kubenswrapper[4727]: E0929 10:39:11.933867 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" podUID="dd80e64b-eafd-40c9-8f06-6233708a7ee5" Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.267854 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03aebbc4-3f74-46f5-84b9-c0b7935569b7","Type":"ContainerStarted","Data":"9684e84219ccd0abdc76e486a7541f648e284063a4970f063b2f930748080083"} Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.271292 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerStarted","Data":"b910c6dbfec64243cf3bf137a89f4a36681499fae1b69cafb5f2697e92dfcdee"} Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.457552 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.468047 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.474883 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.659247 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.888135 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 29 10:39:12 crc kubenswrapper[4727]: I0929 10:39:12.910946 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4"] Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.097421 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 29 10:39:13 crc kubenswrapper[4727]: W0929 10:39:13.261273 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac9728f6_fdb6_47ba_a17c_4220301c2a88.slice/crio-cff764fa78c139a13fc9d16d1da9d38549e595140a22ed004ac7fc3d355522af WatchSource:0}: Error finding container cff764fa78c139a13fc9d16d1da9d38549e595140a22ed004ac7fc3d355522af: Status 404 returned error can't find the container with id cff764fa78c139a13fc9d16d1da9d38549e595140a22ed004ac7fc3d355522af Sep 29 10:39:13 crc kubenswrapper[4727]: W0929 10:39:13.272712 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36e70287_3b92_41e6_a056_fb29d1e03772.slice/crio-c77a189a1780eea703a8cb35c9e212df3336667ada028447690835c5c571b781 WatchSource:0}: Error finding container c77a189a1780eea703a8cb35c9e212df3336667ada028447690835c5c571b781: Status 404 returned error can't find the container with id c77a189a1780eea703a8cb35c9e212df3336667ada028447690835c5c571b781 Sep 29 10:39:13 crc kubenswrapper[4727]: W0929 10:39:13.277769 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bb550f2_83cb_472e_9e1e_b5da4779b42d.slice/crio-30d49f3abdee3455d22887b55bc7d468bcf2126438ebff30f6ea51ef52434bc7 WatchSource:0}: Error finding container 30d49f3abdee3455d22887b55bc7d468bcf2126438ebff30f6ea51ef52434bc7: Status 404 returned error can't find the container with id 30d49f3abdee3455d22887b55bc7d468bcf2126438ebff30f6ea51ef52434bc7 Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.284135 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerStarted","Data":"f647d4c0c82889d7d4443a6f86c3f021bd40f960896fda7799a1b92d950e3615"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.296425 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d25415e-3c8d-4cd3-ab20-83a815a5e39f","Type":"ContainerStarted","Data":"082775f8cc712434038226ae10618ebef4c9018d6f6cef4138469d7a28f7cc46"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.298282 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" event={"ID":"be36e2ed-2475-4b53-8d00-ae2b616cfac0","Type":"ContainerDied","Data":"e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.298322 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4e8943765b64a816d0f5a25c02440e66eb03639506caab6ab50ab3fbf6d1cae" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.301703 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4" event={"ID":"36e70287-3b92-41e6-a056-fb29d1e03772","Type":"ContainerStarted","Data":"c77a189a1780eea703a8cb35c9e212df3336667ada028447690835c5c571b781"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.305505 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8f2d7f61-fe27-4af2-8ef4-05d247fde09d","Type":"ContainerStarted","Data":"d6f370a63c10ba87d6800e45c343f7c4ee8a9f6c3f474cd73c6ae66eaf17afb8"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.318177 4727 generic.go:334] "Generic (PLEG): container finished" podID="39f8267e-103f-4752-847e-d82db8663bab" containerID="df8ff4fabf46f6ff5add17572900fe3032dc45eddb28f87fe5d681596948309c" exitCode=0 Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.318276 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" event={"ID":"39f8267e-103f-4752-847e-d82db8663bab","Type":"ContainerDied","Data":"df8ff4fabf46f6ff5add17572900fe3032dc45eddb28f87fe5d681596948309c"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.318314 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" event={"ID":"39f8267e-103f-4752-847e-d82db8663bab","Type":"ContainerStarted","Data":"887e68baacb6ae198440b70186b61ade9d63b69381af568dc810d871b5d7ac3f"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.323745 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ac9728f6-fdb6-47ba-a17c-4220301c2a88","Type":"ContainerStarted","Data":"cff764fa78c139a13fc9d16d1da9d38549e595140a22ed004ac7fc3d355522af"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.332167 4727 generic.go:334] "Generic (PLEG): container finished" podID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerID="0c8fa8d5244b4a6270882b3711ebe4ed72081927b2979056ce4d78922fc2063d" exitCode=0 Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.332233 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" event={"ID":"bd664253-6a4b-4b3f-baba-ad5ef08b8485","Type":"ContainerDied","Data":"0c8fa8d5244b4a6270882b3711ebe4ed72081927b2979056ce4d78922fc2063d"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.334801 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" event={"ID":"dd80e64b-eafd-40c9-8f06-6233708a7ee5","Type":"ContainerDied","Data":"8272a424752ddc856a90576b23566cdca9b1b83a08ee7a4237511cd69e19bc2a"} Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.334831 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8272a424752ddc856a90576b23566cdca9b1b83a08ee7a4237511cd69e19bc2a" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.356779 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.367923 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.395523 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.403175 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.408882 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.409171 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dllcf" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.410992 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.411957 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.421171 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config\") pod \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.421264 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58jkg\" (UniqueName: \"kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg\") pod \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\" (UID: \"be36e2ed-2475-4b53-8d00-ae2b616cfac0\") " Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.421381 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config\") pod \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.421409 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl2xx\" (UniqueName: \"kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx\") pod \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.421428 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc\") pod \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\" (UID: \"dd80e64b-eafd-40c9-8f06-6233708a7ee5\") " Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.423938 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd80e64b-eafd-40c9-8f06-6233708a7ee5" (UID: "dd80e64b-eafd-40c9-8f06-6233708a7ee5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.424573 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config" (OuterVolumeSpecName: "config") pod "dd80e64b-eafd-40c9-8f06-6233708a7ee5" (UID: "dd80e64b-eafd-40c9-8f06-6233708a7ee5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.425267 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config" (OuterVolumeSpecName: "config") pod "be36e2ed-2475-4b53-8d00-ae2b616cfac0" (UID: "be36e2ed-2475-4b53-8d00-ae2b616cfac0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.433490 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.435650 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx" (OuterVolumeSpecName: "kube-api-access-jl2xx") pod "dd80e64b-eafd-40c9-8f06-6233708a7ee5" (UID: "dd80e64b-eafd-40c9-8f06-6233708a7ee5"). InnerVolumeSpecName "kube-api-access-jl2xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.443104 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg" (OuterVolumeSpecName: "kube-api-access-58jkg") pod "be36e2ed-2475-4b53-8d00-ae2b616cfac0" (UID: "be36e2ed-2475-4b53-8d00-ae2b616cfac0"). InnerVolumeSpecName "kube-api-access-58jkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523607 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523683 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523730 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p87qm\" (UniqueName: \"kubernetes.io/projected/13550429-27b2-4d79-8a60-447c329c77f6-kube-api-access-p87qm\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523783 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13550429-27b2-4d79-8a60-447c329c77f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523841 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523863 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523902 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.523937 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.524001 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58jkg\" (UniqueName: \"kubernetes.io/projected/be36e2ed-2475-4b53-8d00-ae2b616cfac0-kube-api-access-58jkg\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.524016 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.524027 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl2xx\" (UniqueName: \"kubernetes.io/projected/dd80e64b-eafd-40c9-8f06-6233708a7ee5-kube-api-access-jl2xx\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.524041 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd80e64b-eafd-40c9-8f06-6233708a7ee5-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.524052 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be36e2ed-2475-4b53-8d00-ae2b616cfac0-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626248 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13550429-27b2-4d79-8a60-447c329c77f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626403 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626496 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626565 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626634 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626674 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13550429-27b2-4d79-8a60-447c329c77f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626715 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626760 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.626836 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p87qm\" (UniqueName: \"kubernetes.io/projected/13550429-27b2-4d79-8a60-447c329c77f6-kube-api-access-p87qm\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.627601 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.627792 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.627882 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13550429-27b2-4d79-8a60-447c329c77f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.631288 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.634376 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.635998 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13550429-27b2-4d79-8a60-447c329c77f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.646029 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p87qm\" (UniqueName: \"kubernetes.io/projected/13550429-27b2-4d79-8a60-447c329c77f6-kube-api-access-p87qm\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.661857 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"13550429-27b2-4d79-8a60-447c329c77f6\") " pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.687182 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fhhrr"] Sep 29 10:39:13 crc kubenswrapper[4727]: I0929 10:39:13.731137 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:13 crc kubenswrapper[4727]: W0929 10:39:13.747129 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cb18806_a80a_491a_8ade_6371af8b54a9.slice/crio-0b0e560a9f86c5b8e9f95cb1bb0b42466990a24bff331e8d1ceff35c5ed32278 WatchSource:0}: Error finding container 0b0e560a9f86c5b8e9f95cb1bb0b42466990a24bff331e8d1ceff35c5ed32278: Status 404 returned error can't find the container with id 0b0e560a9f86c5b8e9f95cb1bb0b42466990a24bff331e8d1ceff35c5ed32278 Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.343891 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhhrr" event={"ID":"7cb18806-a80a-491a-8ade-6371af8b54a9","Type":"ContainerStarted","Data":"0b0e560a9f86c5b8e9f95cb1bb0b42466990a24bff331e8d1ceff35c5ed32278"} Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.346101 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n98xj" Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.347507 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4bb550f2-83cb-472e-9e1e-b5da4779b42d","Type":"ContainerStarted","Data":"30d49f3abdee3455d22887b55bc7d468bcf2126438ebff30f6ea51ef52434bc7"} Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.347615 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rbhdz" Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.418867 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.432733 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n98xj"] Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.454743 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:39:14 crc kubenswrapper[4727]: I0929 10:39:14.459893 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rbhdz"] Sep 29 10:39:15 crc kubenswrapper[4727]: I0929 10:39:15.117895 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be36e2ed-2475-4b53-8d00-ae2b616cfac0" path="/var/lib/kubelet/pods/be36e2ed-2475-4b53-8d00-ae2b616cfac0/volumes" Sep 29 10:39:15 crc kubenswrapper[4727]: I0929 10:39:15.118323 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd80e64b-eafd-40c9-8f06-6233708a7ee5" path="/var/lib/kubelet/pods/dd80e64b-eafd-40c9-8f06-6233708a7ee5/volumes" Sep 29 10:39:17 crc kubenswrapper[4727]: I0929 10:39:17.031884 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 29 10:39:19 crc kubenswrapper[4727]: I0929 10:39:19.383325 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13550429-27b2-4d79-8a60-447c329c77f6","Type":"ContainerStarted","Data":"7f00a5eb07f8fc72435cc09cb604119b4ba52064dafcd8bffacfd6094fcbe267"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.391305 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d25415e-3c8d-4cd3-ab20-83a815a5e39f","Type":"ContainerStarted","Data":"7eb61495329c6a244ef5e2b7dbd516768708b9481edecfdd326a85ba74f2bdf7"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.402243 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ac9728f6-fdb6-47ba-a17c-4220301c2a88","Type":"ContainerStarted","Data":"77c1887583489b789f1d31551e2b3fc3d5002d7daa24703c3dd6c010e7508529"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.404820 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" event={"ID":"bd664253-6a4b-4b3f-baba-ad5ef08b8485","Type":"ContainerStarted","Data":"fe1f33f8efaffd056c3223cd4a75a4b26281760caef8209334f15abd2d836747"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.405276 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.407159 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8f2d7f61-fe27-4af2-8ef4-05d247fde09d","Type":"ContainerStarted","Data":"d61f3495aca02dd3394796727bcc36678155ead553272ab4ce3efdbbb30c3e2a"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.407209 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.411980 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhhrr" event={"ID":"7cb18806-a80a-491a-8ade-6371af8b54a9","Type":"ContainerStarted","Data":"3f8a9059ea8f1cb2d9c9bd53e1444c85a5d481949a6ab9e79e9715442f1ea55d"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.417044 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerStarted","Data":"1f730413d55de9f8401a7955c284e7d886a5fe384f645986bcdcbe43612ec7b6"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.419770 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" event={"ID":"39f8267e-103f-4752-847e-d82db8663bab","Type":"ContainerStarted","Data":"9ef50efa179702acd5bc874a74c3f042103c767ca1c0edd0b0039a4c4dc646aa"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.420474 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.423596 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03aebbc4-3f74-46f5-84b9-c0b7935569b7","Type":"ContainerStarted","Data":"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.424413 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.426100 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4bb550f2-83cb-472e-9e1e-b5da4779b42d","Type":"ContainerStarted","Data":"5aa0d0c1aaeb914ddb3ac299b98ebeee102e4eec77fb2d42a8a5f0ec1198325e"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.429918 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerStarted","Data":"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864"} Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.438265 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.871753609 podStartE2EDuration="17.438244709s" podCreationTimestamp="2025-09-29 10:39:03 +0000 UTC" firstStartedPulling="2025-09-29 10:39:12.739025301 +0000 UTC m=+1022.912338663" lastFinishedPulling="2025-09-29 10:39:19.305516391 +0000 UTC m=+1029.478829763" observedRunningTime="2025-09-29 10:39:20.435649373 +0000 UTC m=+1030.608962735" watchObservedRunningTime="2025-09-29 10:39:20.438244709 +0000 UTC m=+1030.611558071" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.480157 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" podStartSLOduration=16.112287711 podStartE2EDuration="21.480138494s" podCreationTimestamp="2025-09-29 10:38:59 +0000 UTC" firstStartedPulling="2025-09-29 10:39:06.670005293 +0000 UTC m=+1016.843318655" lastFinishedPulling="2025-09-29 10:39:12.037856076 +0000 UTC m=+1022.211169438" observedRunningTime="2025-09-29 10:39:20.476212703 +0000 UTC m=+1030.649526065" watchObservedRunningTime="2025-09-29 10:39:20.480138494 +0000 UTC m=+1030.653451866" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.598505 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" podStartSLOduration=21.598485658 podStartE2EDuration="21.598485658s" podCreationTimestamp="2025-09-29 10:38:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:20.562575757 +0000 UTC m=+1030.735889119" watchObservedRunningTime="2025-09-29 10:39:20.598485658 +0000 UTC m=+1030.771799020" Sep 29 10:39:20 crc kubenswrapper[4727]: I0929 10:39:20.599165 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=7.653540069 podStartE2EDuration="15.599160405s" podCreationTimestamp="2025-09-29 10:39:05 +0000 UTC" firstStartedPulling="2025-09-29 10:39:11.923815282 +0000 UTC m=+1022.097128644" lastFinishedPulling="2025-09-29 10:39:19.869435618 +0000 UTC m=+1030.042748980" observedRunningTime="2025-09-29 10:39:20.549512612 +0000 UTC m=+1030.722825974" watchObservedRunningTime="2025-09-29 10:39:20.599160405 +0000 UTC m=+1030.772473767" Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.441119 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4" event={"ID":"36e70287-3b92-41e6-a056-fb29d1e03772","Type":"ContainerStarted","Data":"70e5873b6a3e2577a5756c0ddec813c6321031e3d5ea18b2204768ac7d455973"} Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.441610 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-x86c4" Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.443263 4727 generic.go:334] "Generic (PLEG): container finished" podID="7cb18806-a80a-491a-8ade-6371af8b54a9" containerID="3f8a9059ea8f1cb2d9c9bd53e1444c85a5d481949a6ab9e79e9715442f1ea55d" exitCode=0 Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.443383 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhhrr" event={"ID":"7cb18806-a80a-491a-8ade-6371af8b54a9","Type":"ContainerDied","Data":"3f8a9059ea8f1cb2d9c9bd53e1444c85a5d481949a6ab9e79e9715442f1ea55d"} Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.446238 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13550429-27b2-4d79-8a60-447c329c77f6","Type":"ContainerStarted","Data":"e2d49dc515863baa4ffa3c70886d5f8025704d7f77c6d8c4106b816778bfd88f"} Sep 29 10:39:21 crc kubenswrapper[4727]: I0929 10:39:21.469711 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-x86c4" podStartSLOduration=4.871511549 podStartE2EDuration="11.469689882s" podCreationTimestamp="2025-09-29 10:39:10 +0000 UTC" firstStartedPulling="2025-09-29 10:39:13.277275989 +0000 UTC m=+1023.450589351" lastFinishedPulling="2025-09-29 10:39:19.875454322 +0000 UTC m=+1030.048767684" observedRunningTime="2025-09-29 10:39:21.459721947 +0000 UTC m=+1031.633035349" watchObservedRunningTime="2025-09-29 10:39:21.469689882 +0000 UTC m=+1031.643003244" Sep 29 10:39:22 crc kubenswrapper[4727]: I0929 10:39:22.458244 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhhrr" event={"ID":"7cb18806-a80a-491a-8ade-6371af8b54a9","Type":"ContainerStarted","Data":"fbc850d26a93dfc4aa6ddede7be4326c54b2456ed597f42c8296609323b0350d"} Sep 29 10:39:22 crc kubenswrapper[4727]: I0929 10:39:22.458633 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhhrr" event={"ID":"7cb18806-a80a-491a-8ade-6371af8b54a9","Type":"ContainerStarted","Data":"015aa664730c189e5dd4e60c3c2dfd63a46c1f0484f112ac8fdaad7b236eab3e"} Sep 29 10:39:22 crc kubenswrapper[4727]: I0929 10:39:22.459489 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:22 crc kubenswrapper[4727]: I0929 10:39:22.459510 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.165116 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.182618 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fhhrr" podStartSLOduration=9.061566147 podStartE2EDuration="15.182598898s" podCreationTimestamp="2025-09-29 10:39:10 +0000 UTC" firstStartedPulling="2025-09-29 10:39:13.753822036 +0000 UTC m=+1023.927135398" lastFinishedPulling="2025-09-29 10:39:19.874854787 +0000 UTC m=+1030.048168149" observedRunningTime="2025-09-29 10:39:22.483049671 +0000 UTC m=+1032.656363033" watchObservedRunningTime="2025-09-29 10:39:25.182598898 +0000 UTC m=+1035.355912260" Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.232755 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.233047 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="dnsmasq-dns" containerID="cri-o://fe1f33f8efaffd056c3223cd4a75a4b26281760caef8209334f15abd2d836747" gracePeriod=10 Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.236133 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.478782 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d25415e-3c8d-4cd3-ab20-83a815a5e39f","Type":"ContainerDied","Data":"7eb61495329c6a244ef5e2b7dbd516768708b9481edecfdd326a85ba74f2bdf7"} Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.478863 4727 generic.go:334] "Generic (PLEG): container finished" podID="2d25415e-3c8d-4cd3-ab20-83a815a5e39f" containerID="7eb61495329c6a244ef5e2b7dbd516768708b9481edecfdd326a85ba74f2bdf7" exitCode=0 Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.480599 4727 generic.go:334] "Generic (PLEG): container finished" podID="ac9728f6-fdb6-47ba-a17c-4220301c2a88" containerID="77c1887583489b789f1d31551e2b3fc3d5002d7daa24703c3dd6c010e7508529" exitCode=0 Sep 29 10:39:25 crc kubenswrapper[4727]: I0929 10:39:25.480631 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ac9728f6-fdb6-47ba-a17c-4220301c2a88","Type":"ContainerDied","Data":"77c1887583489b789f1d31551e2b3fc3d5002d7daa24703c3dd6c010e7508529"} Sep 29 10:39:26 crc kubenswrapper[4727]: I0929 10:39:26.153680 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 29 10:39:26 crc kubenswrapper[4727]: I0929 10:39:26.489022 4727 generic.go:334] "Generic (PLEG): container finished" podID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerID="fe1f33f8efaffd056c3223cd4a75a4b26281760caef8209334f15abd2d836747" exitCode=0 Sep 29 10:39:26 crc kubenswrapper[4727]: I0929 10:39:26.489067 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" event={"ID":"bd664253-6a4b-4b3f-baba-ad5ef08b8485","Type":"ContainerDied","Data":"fe1f33f8efaffd056c3223cd4a75a4b26281760caef8209334f15abd2d836747"} Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.165106 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.343564 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config\") pod \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.343632 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b75fl\" (UniqueName: \"kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl\") pod \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.343664 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc\") pod \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\" (UID: \"bd664253-6a4b-4b3f-baba-ad5ef08b8485\") " Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.349488 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl" (OuterVolumeSpecName: "kube-api-access-b75fl") pod "bd664253-6a4b-4b3f-baba-ad5ef08b8485" (UID: "bd664253-6a4b-4b3f-baba-ad5ef08b8485"). InnerVolumeSpecName "kube-api-access-b75fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.383098 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config" (OuterVolumeSpecName: "config") pod "bd664253-6a4b-4b3f-baba-ad5ef08b8485" (UID: "bd664253-6a4b-4b3f-baba-ad5ef08b8485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.383783 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd664253-6a4b-4b3f-baba-ad5ef08b8485" (UID: "bd664253-6a4b-4b3f-baba-ad5ef08b8485"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.445603 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.445642 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b75fl\" (UniqueName: \"kubernetes.io/projected/bd664253-6a4b-4b3f-baba-ad5ef08b8485-kube-api-access-b75fl\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.445655 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd664253-6a4b-4b3f-baba-ad5ef08b8485-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.497492 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2d25415e-3c8d-4cd3-ab20-83a815a5e39f","Type":"ContainerStarted","Data":"e1b4b75a3805042ebbcc05c455b338c9ac35affc9023e9ea3dbe9a6bbe2a54e7"} Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.500941 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ac9728f6-fdb6-47ba-a17c-4220301c2a88","Type":"ContainerStarted","Data":"0a1b09858ccfaa4de20cb7e1b0794c7db9813f525935144278b35af6923ae924"} Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.502976 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" event={"ID":"bd664253-6a4b-4b3f-baba-ad5ef08b8485","Type":"ContainerDied","Data":"05d92e582a33338b3ad469960381f09b12f79e8fd14ae3054919c1978b0fd141"} Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.503035 4727 scope.go:117] "RemoveContainer" containerID="fe1f33f8efaffd056c3223cd4a75a4b26281760caef8209334f15abd2d836747" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.503166 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-w6pdj" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.521049 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.91298901 podStartE2EDuration="25.521025366s" podCreationTimestamp="2025-09-29 10:39:02 +0000 UTC" firstStartedPulling="2025-09-29 10:39:13.265466896 +0000 UTC m=+1023.438780258" lastFinishedPulling="2025-09-29 10:39:19.873503262 +0000 UTC m=+1030.046816614" observedRunningTime="2025-09-29 10:39:27.517041264 +0000 UTC m=+1037.690354626" watchObservedRunningTime="2025-09-29 10:39:27.521025366 +0000 UTC m=+1037.694338728" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.536204 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.536387 4727 scope.go:117] "RemoveContainer" containerID="0c8fa8d5244b4a6270882b3711ebe4ed72081927b2979056ce4d78922fc2063d" Sep 29 10:39:27 crc kubenswrapper[4727]: I0929 10:39:27.542583 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-w6pdj"] Sep 29 10:39:28 crc kubenswrapper[4727]: I0929 10:39:28.539333 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.135846735 podStartE2EDuration="27.539313862s" podCreationTimestamp="2025-09-29 10:39:01 +0000 UTC" firstStartedPulling="2025-09-29 10:39:12.472049967 +0000 UTC m=+1022.645363319" lastFinishedPulling="2025-09-29 10:39:19.875517084 +0000 UTC m=+1030.048830446" observedRunningTime="2025-09-29 10:39:28.53300783 +0000 UTC m=+1038.706321202" watchObservedRunningTime="2025-09-29 10:39:28.539313862 +0000 UTC m=+1038.712627224" Sep 29 10:39:29 crc kubenswrapper[4727]: I0929 10:39:29.116991 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" path="/var/lib/kubelet/pods/bd664253-6a4b-4b3f-baba-ad5ef08b8485/volumes" Sep 29 10:39:29 crc kubenswrapper[4727]: I0929 10:39:29.256530 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 29 10:39:32 crc kubenswrapper[4727]: I0929 10:39:32.842161 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 29 10:39:32 crc kubenswrapper[4727]: I0929 10:39:32.842829 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 29 10:39:33 crc kubenswrapper[4727]: E0929 10:39:33.150666 4727 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.153:37180->38.102.83.153:44021: write tcp 38.102.83.153:37180->38.102.83.153:44021: write: broken pipe Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.546959 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4bb550f2-83cb-472e-9e1e-b5da4779b42d","Type":"ContainerStarted","Data":"5c01a17b87560bd8f1287a22ba2a148d71adc08757d02439144e01df016ce3b5"} Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.548497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13550429-27b2-4d79-8a60-447c329c77f6","Type":"ContainerStarted","Data":"dd1c64321c9ad95f7d651360d5c54d31e5959610ce4bb718432607712f0031a4"} Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.571538 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=6.431561661 podStartE2EDuration="25.57152158s" podCreationTimestamp="2025-09-29 10:39:08 +0000 UTC" firstStartedPulling="2025-09-29 10:39:13.286871935 +0000 UTC m=+1023.460185297" lastFinishedPulling="2025-09-29 10:39:32.426831854 +0000 UTC m=+1042.600145216" observedRunningTime="2025-09-29 10:39:33.567594939 +0000 UTC m=+1043.740908301" watchObservedRunningTime="2025-09-29 10:39:33.57152158 +0000 UTC m=+1043.744834942" Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.590759 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.473909484 podStartE2EDuration="21.590743522s" podCreationTimestamp="2025-09-29 10:39:12 +0000 UTC" firstStartedPulling="2025-09-29 10:39:19.333318024 +0000 UTC m=+1029.506631386" lastFinishedPulling="2025-09-29 10:39:32.450152052 +0000 UTC m=+1042.623465424" observedRunningTime="2025-09-29 10:39:33.583969479 +0000 UTC m=+1043.757282841" watchObservedRunningTime="2025-09-29 10:39:33.590743522 +0000 UTC m=+1043.764056884" Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.732297 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.925582 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:33 crc kubenswrapper[4727]: I0929 10:39:33.925645 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:34 crc kubenswrapper[4727]: I0929 10:39:34.527114 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:34 crc kubenswrapper[4727]: I0929 10:39:34.731416 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:34 crc kubenswrapper[4727]: I0929 10:39:34.780744 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:34 crc kubenswrapper[4727]: I0929 10:39:34.891463 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 29 10:39:34 crc kubenswrapper[4727]: I0929 10:39:34.946807 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.597195 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.826917 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:35 crc kubenswrapper[4727]: E0929 10:39:35.829540 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="dnsmasq-dns" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.829574 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="dnsmasq-dns" Sep 29 10:39:35 crc kubenswrapper[4727]: E0929 10:39:35.829609 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="init" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.829617 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="init" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.829766 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd664253-6a4b-4b3f-baba-ad5ef08b8485" containerName="dnsmasq-dns" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.830618 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.836787 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.841708 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.884123 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.884199 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.884229 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.884257 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc692\" (UniqueName: \"kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.939453 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-rvsjj"] Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.940422 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.945616 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.949433 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rvsjj"] Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986244 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc692\" (UniqueName: \"kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986312 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7a1be1-e75b-47f3-8b78-48f30fa37940-config\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986450 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovn-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986482 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovs-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986505 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsl9p\" (UniqueName: \"kubernetes.io/projected/ad7a1be1-e75b-47f3-8b78-48f30fa37940-kube-api-access-gsl9p\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986528 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-combined-ca-bundle\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986548 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986598 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986622 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.986642 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.987674 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.987701 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:35 crc kubenswrapper[4727]: I0929 10:39:35.987683 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.011394 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc692\" (UniqueName: \"kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692\") pod \"dnsmasq-dns-7f896c8c65-rsgwq\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089059 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovn-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089622 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovs-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsl9p\" (UniqueName: \"kubernetes.io/projected/ad7a1be1-e75b-47f3-8b78-48f30fa37940-kube-api-access-gsl9p\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089817 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-combined-ca-bundle\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089965 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.090147 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7a1be1-e75b-47f3-8b78-48f30fa37940-config\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089428 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovn-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.091055 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7a1be1-e75b-47f3-8b78-48f30fa37940-config\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.089709 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ad7a1be1-e75b-47f3-8b78-48f30fa37940-ovs-rundir\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.096207 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-combined-ca-bundle\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.103229 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad7a1be1-e75b-47f3-8b78-48f30fa37940-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.121211 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsl9p\" (UniqueName: \"kubernetes.io/projected/ad7a1be1-e75b-47f3-8b78-48f30fa37940-kube-api-access-gsl9p\") pod \"ovn-controller-metrics-rvsjj\" (UID: \"ad7a1be1-e75b-47f3-8b78-48f30fa37940\") " pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.153707 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.162539 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.217041 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-pkrp9"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.220539 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.249871 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-pkrp9"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.257902 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rvsjj" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.300598 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.301924 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.302076 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsmhh\" (UniqueName: \"kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.302238 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.396004 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-pkrp9"] Sep 29 10:39:36 crc kubenswrapper[4727]: E0929 10:39:36.396709 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-nsmhh ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" podUID="5b167619-b7df-403c-8316-c907d096f168" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.406226 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.406298 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.406623 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsmhh\" (UniqueName: \"kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.406742 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.407305 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.407820 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.407854 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.429217 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.434124 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.441235 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.461722 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsmhh\" (UniqueName: \"kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh\") pod \"dnsmasq-dns-6c89d5d749-pkrp9\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.461793 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.510412 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.510558 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crl2q\" (UniqueName: \"kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.510712 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.510786 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.510824 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.527642 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.570535 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.585609 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.595791 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612042 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612084 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crl2q\" (UniqueName: \"kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612144 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612187 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612207 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.612997 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.613586 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.616384 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.616624 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.632545 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crl2q\" (UniqueName: \"kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q\") pod \"dnsmasq-dns-698758b865-sknk6\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.713649 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc\") pod \"5b167619-b7df-403c-8316-c907d096f168\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.713787 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsmhh\" (UniqueName: \"kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh\") pod \"5b167619-b7df-403c-8316-c907d096f168\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.713866 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb\") pod \"5b167619-b7df-403c-8316-c907d096f168\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.713961 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config\") pod \"5b167619-b7df-403c-8316-c907d096f168\" (UID: \"5b167619-b7df-403c-8316-c907d096f168\") " Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.715607 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b167619-b7df-403c-8316-c907d096f168" (UID: "5b167619-b7df-403c-8316-c907d096f168"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.717049 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b167619-b7df-403c-8316-c907d096f168" (UID: "5b167619-b7df-403c-8316-c907d096f168"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.719092 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config" (OuterVolumeSpecName: "config") pod "5b167619-b7df-403c-8316-c907d096f168" (UID: "5b167619-b7df-403c-8316-c907d096f168"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.719418 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh" (OuterVolumeSpecName: "kube-api-access-nsmhh") pod "5b167619-b7df-403c-8316-c907d096f168" (UID: "5b167619-b7df-403c-8316-c907d096f168"). InnerVolumeSpecName "kube-api-access-nsmhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.747533 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:36 crc kubenswrapper[4727]: W0929 10:39:36.750676 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a86bfd1_a586_4f2b_ad59_50847cc0698d.slice/crio-2ce82e740a7e62f7c3c9ab22caf519daf5079a74fa5a75c5f7dbbd87208f4d98 WatchSource:0}: Error finding container 2ce82e740a7e62f7c3c9ab22caf519daf5079a74fa5a75c5f7dbbd87208f4d98: Status 404 returned error can't find the container with id 2ce82e740a7e62f7c3c9ab22caf519daf5079a74fa5a75c5f7dbbd87208f4d98 Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.763414 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.814183 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rvsjj"] Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.815900 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.815933 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsmhh\" (UniqueName: \"kubernetes.io/projected/5b167619-b7df-403c-8316-c907d096f168-kube-api-access-nsmhh\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.815947 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:36 crc kubenswrapper[4727]: I0929 10:39:36.815959 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b167619-b7df-403c-8316-c907d096f168-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.184371 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:39:37 crc kubenswrapper[4727]: W0929 10:39:37.203460 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4460bc15_e725_4559_ab19_344e52088e27.slice/crio-0a94ecc3d55ac0e9bce676200c8533aa624afb67a077454d163e2de95ea554c9 WatchSource:0}: Error finding container 0a94ecc3d55ac0e9bce676200c8533aa624afb67a077454d163e2de95ea554c9: Status 404 returned error can't find the container with id 0a94ecc3d55ac0e9bce676200c8533aa624afb67a077454d163e2de95ea554c9 Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.281681 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.288736 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.291617 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.291615 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.291945 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.292211 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5kkrr" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.304112 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.425585 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.425637 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-cache\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.425669 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-lock\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.425691 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.425709 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl9cd\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-kube-api-access-xl9cd\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528064 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528130 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-cache\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528180 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-lock\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528205 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528231 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl9cd\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-kube-api-access-xl9cd\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.528496 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: E0929 10:39:37.528815 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:37 crc kubenswrapper[4727]: E0929 10:39:37.528849 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:37 crc kubenswrapper[4727]: E0929 10:39:37.528914 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:39:38.028893622 +0000 UTC m=+1048.202206984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.529364 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-lock\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.529402 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7a6a9dee-ca84-4c2d-abc9-613bccd90764-cache\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.549897 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl9cd\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-kube-api-access-xl9cd\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.550425 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.581200 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sknk6" event={"ID":"4460bc15-e725-4559-ab19-344e52088e27","Type":"ContainerStarted","Data":"0a94ecc3d55ac0e9bce676200c8533aa624afb67a077454d163e2de95ea554c9"} Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.583752 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rvsjj" event={"ID":"ad7a1be1-e75b-47f3-8b78-48f30fa37940","Type":"ContainerStarted","Data":"4a25e6f4b605da3c1920abd4aa94d5f220b9cc783ec2aa645145a168d41caa21"} Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.583895 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rvsjj" event={"ID":"ad7a1be1-e75b-47f3-8b78-48f30fa37940","Type":"ContainerStarted","Data":"1baaa883a004453c8fc23e297bf4a9cc4b19ea83ddfa243712046bbb241a34fb"} Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.587021 4727 generic.go:334] "Generic (PLEG): container finished" podID="9a86bfd1-a586-4f2b-ad59-50847cc0698d" containerID="654ccbc0aed0c81b760005bb20003f2568c34f81f575c586af07af6310e364fa" exitCode=0 Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.587190 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" event={"ID":"9a86bfd1-a586-4f2b-ad59-50847cc0698d","Type":"ContainerDied","Data":"654ccbc0aed0c81b760005bb20003f2568c34f81f575c586af07af6310e364fa"} Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.587316 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" event={"ID":"9a86bfd1-a586-4f2b-ad59-50847cc0698d","Type":"ContainerStarted","Data":"2ce82e740a7e62f7c3c9ab22caf519daf5079a74fa5a75c5f7dbbd87208f4d98"} Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.587594 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-pkrp9" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.612611 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-rvsjj" podStartSLOduration=2.612576447 podStartE2EDuration="2.612576447s" podCreationTimestamp="2025-09-29 10:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:37.602372486 +0000 UTC m=+1047.775685868" watchObservedRunningTime="2025-09-29 10:39:37.612576447 +0000 UTC m=+1047.785889809" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.651380 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.675745 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-pkrp9"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.690204 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-pkrp9"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.857476 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.861459 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.871228 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.871422 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.871575 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7q9v9" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.872671 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.871436 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.914826 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.935142 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config\") pod \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.936055 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc\") pod \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.936294 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc692\" (UniqueName: \"kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692\") pod \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.936450 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb\") pod \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\" (UID: \"9a86bfd1-a586-4f2b-ad59-50847cc0698d\") " Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.936891 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-scripts\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937049 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937179 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-config\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937289 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937472 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntj96\" (UniqueName: \"kubernetes.io/projected/2db0e30f-219c-4539-8f02-3f64f9f23a50-kube-api-access-ntj96\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937618 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.937766 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.942326 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692" (OuterVolumeSpecName: "kube-api-access-nc692") pod "9a86bfd1-a586-4f2b-ad59-50847cc0698d" (UID: "9a86bfd1-a586-4f2b-ad59-50847cc0698d"). InnerVolumeSpecName "kube-api-access-nc692". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.954909 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a86bfd1-a586-4f2b-ad59-50847cc0698d" (UID: "9a86bfd1-a586-4f2b-ad59-50847cc0698d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.955207 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config" (OuterVolumeSpecName: "config") pod "9a86bfd1-a586-4f2b-ad59-50847cc0698d" (UID: "9a86bfd1-a586-4f2b-ad59-50847cc0698d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:37 crc kubenswrapper[4727]: I0929 10:39:37.959452 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a86bfd1-a586-4f2b-ad59-50847cc0698d" (UID: "9a86bfd1-a586-4f2b-ad59-50847cc0698d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.040293 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.040552 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.040701 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-scripts\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.040825 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.040949 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-config\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041069 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041360 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041784 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntj96\" (UniqueName: \"kubernetes.io/projected/2db0e30f-219c-4539-8f02-3f64f9f23a50-kube-api-access-ntj96\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041970 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.042088 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.042174 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc692\" (UniqueName: \"kubernetes.io/projected/9a86bfd1-a586-4f2b-ad59-50847cc0698d-kube-api-access-nc692\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.042260 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a86bfd1-a586-4f2b-ad59-50847cc0698d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041256 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041613 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-scripts\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: E0929 10:39:38.041453 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:38 crc kubenswrapper[4727]: E0929 10:39:38.042650 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:38 crc kubenswrapper[4727]: E0929 10:39:38.042768 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:39:39.042745344 +0000 UTC m=+1049.216058706 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.041735 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db0e30f-219c-4539-8f02-3f64f9f23a50-config\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.045211 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.045217 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.046180 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e30f-219c-4539-8f02-3f64f9f23a50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.059232 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntj96\" (UniqueName: \"kubernetes.io/projected/2db0e30f-219c-4539-8f02-3f64f9f23a50-kube-api-access-ntj96\") pod \"ovn-northd-0\" (UID: \"2db0e30f-219c-4539-8f02-3f64f9f23a50\") " pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.180879 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.596436 4727 generic.go:334] "Generic (PLEG): container finished" podID="4460bc15-e725-4559-ab19-344e52088e27" containerID="698dbe80bb166c232574ca754314c3dee83cd6a2acadda942b88789556e54f95" exitCode=0 Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.596517 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sknk6" event={"ID":"4460bc15-e725-4559-ab19-344e52088e27","Type":"ContainerDied","Data":"698dbe80bb166c232574ca754314c3dee83cd6a2acadda942b88789556e54f95"} Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.598460 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" event={"ID":"9a86bfd1-a586-4f2b-ad59-50847cc0698d","Type":"ContainerDied","Data":"2ce82e740a7e62f7c3c9ab22caf519daf5079a74fa5a75c5f7dbbd87208f4d98"} Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.598504 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-rsgwq" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.598511 4727 scope.go:117] "RemoveContainer" containerID="654ccbc0aed0c81b760005bb20003f2568c34f81f575c586af07af6310e364fa" Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.663980 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:38 crc kubenswrapper[4727]: W0929 10:39:38.665779 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2db0e30f_219c_4539_8f02_3f64f9f23a50.slice/crio-e968627f9afbf49f32684b562fa15dd0fe99b31e1de01e695c71d63a90eca68e WatchSource:0}: Error finding container e968627f9afbf49f32684b562fa15dd0fe99b31e1de01e695c71d63a90eca68e: Status 404 returned error can't find the container with id e968627f9afbf49f32684b562fa15dd0fe99b31e1de01e695c71d63a90eca68e Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.670054 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-rsgwq"] Sep 29 10:39:38 crc kubenswrapper[4727]: I0929 10:39:38.676011 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.058972 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:39 crc kubenswrapper[4727]: E0929 10:39:39.059264 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:39 crc kubenswrapper[4727]: E0929 10:39:39.059452 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:39 crc kubenswrapper[4727]: E0929 10:39:39.059518 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:39:41.059499301 +0000 UTC m=+1051.232812663 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.121267 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b167619-b7df-403c-8316-c907d096f168" path="/var/lib/kubelet/pods/5b167619-b7df-403c-8316-c907d096f168/volumes" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.121824 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a86bfd1-a586-4f2b-ad59-50847cc0698d" path="/var/lib/kubelet/pods/9a86bfd1-a586-4f2b-ad59-50847cc0698d/volumes" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.519312 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6657k"] Sep 29 10:39:39 crc kubenswrapper[4727]: E0929 10:39:39.519676 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a86bfd1-a586-4f2b-ad59-50847cc0698d" containerName="init" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.519691 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a86bfd1-a586-4f2b-ad59-50847cc0698d" containerName="init" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.519871 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a86bfd1-a586-4f2b-ad59-50847cc0698d" containerName="init" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.520387 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6657k" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.532808 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6657k"] Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.570395 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlbqc\" (UniqueName: \"kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc\") pod \"glance-db-create-6657k\" (UID: \"86c42eb0-b98f-4f6a-9107-0ece0511ccb2\") " pod="openstack/glance-db-create-6657k" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.614816 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2db0e30f-219c-4539-8f02-3f64f9f23a50","Type":"ContainerStarted","Data":"e968627f9afbf49f32684b562fa15dd0fe99b31e1de01e695c71d63a90eca68e"} Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.672211 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlbqc\" (UniqueName: \"kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc\") pod \"glance-db-create-6657k\" (UID: \"86c42eb0-b98f-4f6a-9107-0ece0511ccb2\") " pod="openstack/glance-db-create-6657k" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.690189 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlbqc\" (UniqueName: \"kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc\") pod \"glance-db-create-6657k\" (UID: \"86c42eb0-b98f-4f6a-9107-0ece0511ccb2\") " pod="openstack/glance-db-create-6657k" Sep 29 10:39:39 crc kubenswrapper[4727]: I0929 10:39:39.844164 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6657k" Sep 29 10:39:40 crc kubenswrapper[4727]: I0929 10:39:40.255908 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6657k"] Sep 29 10:39:40 crc kubenswrapper[4727]: I0929 10:39:40.623769 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6657k" event={"ID":"86c42eb0-b98f-4f6a-9107-0ece0511ccb2","Type":"ContainerStarted","Data":"c3ee4395dcf44277e48702a831532113aa3f4ce8d2a0a309a1367188a61f025b"} Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.092570 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:41 crc kubenswrapper[4727]: E0929 10:39:41.092810 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:41 crc kubenswrapper[4727]: E0929 10:39:41.092857 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:41 crc kubenswrapper[4727]: E0929 10:39:41.092959 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:39:45.092928971 +0000 UTC m=+1055.266242373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.237320 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-k5gvg"] Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.238615 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.241320 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.241523 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.241611 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.255006 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-k5gvg"] Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.266374 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6jpg2"] Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.267658 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: E0929 10:39:41.267681 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-mj2d8 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-mj2d8 ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-k5gvg" podUID="e5e25c7d-7922-4216-9071-a30c01f14c70" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.275834 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-k5gvg"] Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.283238 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6jpg2"] Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.308927 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.308984 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309021 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309085 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj2d8\" (UniqueName: \"kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309115 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309150 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309179 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309207 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309234 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309272 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309301 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309364 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309432 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.309457 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sn76\" (UniqueName: \"kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.410911 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.410962 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.410990 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411041 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411061 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sn76\" (UniqueName: \"kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411091 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411109 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411131 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411178 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj2d8\" (UniqueName: \"kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411198 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411221 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411243 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411264 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411284 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411532 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.411897 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.412809 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.414296 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.414607 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.414750 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.421999 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.422054 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.422063 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.422293 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.422359 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.423465 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.430884 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sn76\" (UniqueName: \"kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76\") pod \"swift-ring-rebalance-6jpg2\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.433249 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj2d8\" (UniqueName: \"kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8\") pod \"swift-ring-rebalance-k5gvg\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.587068 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.633243 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.633979 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6657k" event={"ID":"86c42eb0-b98f-4f6a-9107-0ece0511ccb2","Type":"ContainerStarted","Data":"bb0ad55688a8dddba66eb403888fda93b9303a1392918b9850c51986bf1590b9"} Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.686664 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.817156 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.817244 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.817278 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.817313 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj2d8\" (UniqueName: \"kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818117 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818167 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818192 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf\") pod \"e5e25c7d-7922-4216-9071-a30c01f14c70\" (UID: \"e5e25c7d-7922-4216-9071-a30c01f14c70\") " Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818624 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818684 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.818695 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts" (OuterVolumeSpecName: "scripts") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.819631 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.819650 4727 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5e25c7d-7922-4216-9071-a30c01f14c70-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.819661 4727 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5e25c7d-7922-4216-9071-a30c01f14c70-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.823370 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8" (OuterVolumeSpecName: "kube-api-access-mj2d8") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "kube-api-access-mj2d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.823895 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.832502 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.833476 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5e25c7d-7922-4216-9071-a30c01f14c70" (UID: "e5e25c7d-7922-4216-9071-a30c01f14c70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.921212 4727 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.921257 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.921268 4727 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5e25c7d-7922-4216-9071-a30c01f14c70-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:41 crc kubenswrapper[4727]: I0929 10:39:41.921277 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj2d8\" (UniqueName: \"kubernetes.io/projected/e5e25c7d-7922-4216-9071-a30c01f14c70-kube-api-access-mj2d8\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.005361 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6jpg2"] Sep 29 10:39:42 crc kubenswrapper[4727]: W0929 10:39:42.010907 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83f5e1ea_5026_48dc_a5a7_d6d841b45376.slice/crio-2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750 WatchSource:0}: Error finding container 2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750: Status 404 returned error can't find the container with id 2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750 Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.641863 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6jpg2" event={"ID":"83f5e1ea-5026-48dc-a5a7-d6d841b45376","Type":"ContainerStarted","Data":"2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750"} Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.645556 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sknk6" event={"ID":"4460bc15-e725-4559-ab19-344e52088e27","Type":"ContainerStarted","Data":"b3834598a41238f8a7826bd8e8245b873f2d6bbaad9462770c5c1d0ccc4e3928"} Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.645587 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k5gvg" Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.689027 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-k5gvg"] Sep 29 10:39:42 crc kubenswrapper[4727]: I0929 10:39:42.694131 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-k5gvg"] Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.134433 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e25c7d-7922-4216-9071-a30c01f14c70" path="/var/lib/kubelet/pods/e5e25c7d-7922-4216-9071-a30c01f14c70/volumes" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.657066 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.677802 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-6657k" podStartSLOduration=4.677782806 podStartE2EDuration="4.677782806s" podCreationTimestamp="2025-09-29 10:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:43.674119253 +0000 UTC m=+1053.847432625" watchObservedRunningTime="2025-09-29 10:39:43.677782806 +0000 UTC m=+1053.851096178" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.697326 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-sknk6" podStartSLOduration=7.697300357 podStartE2EDuration="7.697300357s" podCreationTimestamp="2025-09-29 10:39:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:43.692663718 +0000 UTC m=+1053.865977080" watchObservedRunningTime="2025-09-29 10:39:43.697300357 +0000 UTC m=+1053.870613759" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.819964 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nh7m5"] Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.821788 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nh7m5" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.828518 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nh7m5"] Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.863812 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9hjb\" (UniqueName: \"kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb\") pod \"keystone-db-create-nh7m5\" (UID: \"5b7d9089-b633-4493-8d9a-1a98f8febda0\") " pod="openstack/keystone-db-create-nh7m5" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.965845 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9hjb\" (UniqueName: \"kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb\") pod \"keystone-db-create-nh7m5\" (UID: \"5b7d9089-b633-4493-8d9a-1a98f8febda0\") " pod="openstack/keystone-db-create-nh7m5" Sep 29 10:39:43 crc kubenswrapper[4727]: I0929 10:39:43.983723 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9hjb\" (UniqueName: \"kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb\") pod \"keystone-db-create-nh7m5\" (UID: \"5b7d9089-b633-4493-8d9a-1a98f8febda0\") " pod="openstack/keystone-db-create-nh7m5" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.140489 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-9kxvj"] Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.141998 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9kxvj" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.143891 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nh7m5" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.153420 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9kxvj"] Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.270073 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kttm\" (UniqueName: \"kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm\") pod \"placement-db-create-9kxvj\" (UID: \"0b6fcc5d-2000-448c-9749-0318107fd488\") " pod="openstack/placement-db-create-9kxvj" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.371289 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kttm\" (UniqueName: \"kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm\") pod \"placement-db-create-9kxvj\" (UID: \"0b6fcc5d-2000-448c-9749-0318107fd488\") " pod="openstack/placement-db-create-9kxvj" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.390264 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kttm\" (UniqueName: \"kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm\") pod \"placement-db-create-9kxvj\" (UID: \"0b6fcc5d-2000-448c-9749-0318107fd488\") " pod="openstack/placement-db-create-9kxvj" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.464580 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9kxvj" Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.578599 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nh7m5"] Sep 29 10:39:44 crc kubenswrapper[4727]: W0929 10:39:44.588433 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b7d9089_b633_4493_8d9a_1a98f8febda0.slice/crio-0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15 WatchSource:0}: Error finding container 0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15: Status 404 returned error can't find the container with id 0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15 Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.665994 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nh7m5" event={"ID":"5b7d9089-b633-4493-8d9a-1a98f8febda0","Type":"ContainerStarted","Data":"0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15"} Sep 29 10:39:44 crc kubenswrapper[4727]: I0929 10:39:44.888423 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9kxvj"] Sep 29 10:39:44 crc kubenswrapper[4727]: W0929 10:39:44.894264 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b6fcc5d_2000_448c_9749_0318107fd488.slice/crio-85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0 WatchSource:0}: Error finding container 85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0: Status 404 returned error can't find the container with id 85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0 Sep 29 10:39:45 crc kubenswrapper[4727]: I0929 10:39:45.189124 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:45 crc kubenswrapper[4727]: E0929 10:39:45.189383 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:45 crc kubenswrapper[4727]: E0929 10:39:45.189404 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:45 crc kubenswrapper[4727]: E0929 10:39:45.189504 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:39:53.18948673 +0000 UTC m=+1063.362800092 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:45 crc kubenswrapper[4727]: I0929 10:39:45.677574 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9kxvj" event={"ID":"0b6fcc5d-2000-448c-9749-0318107fd488","Type":"ContainerStarted","Data":"85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0"} Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.688554 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9kxvj" event={"ID":"0b6fcc5d-2000-448c-9749-0318107fd488","Type":"ContainerStarted","Data":"ad0359a69955155ef944dce607d0eca027d3283393b5e65d8c13989d73036e8e"} Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.692928 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nh7m5" event={"ID":"5b7d9089-b633-4493-8d9a-1a98f8febda0","Type":"ContainerStarted","Data":"c0080b1618a82670df46877e015798cc5d03f7ae3f3ca6664d8acdf2cf193ca4"} Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.695817 4727 generic.go:334] "Generic (PLEG): container finished" podID="86c42eb0-b98f-4f6a-9107-0ece0511ccb2" containerID="bb0ad55688a8dddba66eb403888fda93b9303a1392918b9850c51986bf1590b9" exitCode=0 Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.695863 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6657k" event={"ID":"86c42eb0-b98f-4f6a-9107-0ece0511ccb2","Type":"ContainerDied","Data":"bb0ad55688a8dddba66eb403888fda93b9303a1392918b9850c51986bf1590b9"} Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.765413 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.851325 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:39:46 crc kubenswrapper[4727]: I0929 10:39:46.858226 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="dnsmasq-dns" containerID="cri-o://9ef50efa179702acd5bc874a74c3f042103c767ca1c0edd0b0039a4c4dc646aa" gracePeriod=10 Sep 29 10:39:47 crc kubenswrapper[4727]: I0929 10:39:47.707353 4727 generic.go:334] "Generic (PLEG): container finished" podID="39f8267e-103f-4752-847e-d82db8663bab" containerID="9ef50efa179702acd5bc874a74c3f042103c767ca1c0edd0b0039a4c4dc646aa" exitCode=0 Sep 29 10:39:47 crc kubenswrapper[4727]: I0929 10:39:47.707387 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" event={"ID":"39f8267e-103f-4752-847e-d82db8663bab","Type":"ContainerDied","Data":"9ef50efa179702acd5bc874a74c3f042103c767ca1c0edd0b0039a4c4dc646aa"} Sep 29 10:39:47 crc kubenswrapper[4727]: I0929 10:39:47.733281 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-9kxvj" podStartSLOduration=3.733260014 podStartE2EDuration="3.733260014s" podCreationTimestamp="2025-09-29 10:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:47.72961686 +0000 UTC m=+1057.902930222" watchObservedRunningTime="2025-09-29 10:39:47.733260014 +0000 UTC m=+1057.906573376" Sep 29 10:39:47 crc kubenswrapper[4727]: I0929 10:39:47.751169 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-nh7m5" podStartSLOduration=4.751143842 podStartE2EDuration="4.751143842s" podCreationTimestamp="2025-09-29 10:39:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:39:47.749624364 +0000 UTC m=+1057.922937726" watchObservedRunningTime="2025-09-29 10:39:47.751143842 +0000 UTC m=+1057.924457204" Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.314550 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6657k" Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.452615 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlbqc\" (UniqueName: \"kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc\") pod \"86c42eb0-b98f-4f6a-9107-0ece0511ccb2\" (UID: \"86c42eb0-b98f-4f6a-9107-0ece0511ccb2\") " Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.460673 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc" (OuterVolumeSpecName: "kube-api-access-vlbqc") pod "86c42eb0-b98f-4f6a-9107-0ece0511ccb2" (UID: "86c42eb0-b98f-4f6a-9107-0ece0511ccb2"). InnerVolumeSpecName "kube-api-access-vlbqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.556931 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlbqc\" (UniqueName: \"kubernetes.io/projected/86c42eb0-b98f-4f6a-9107-0ece0511ccb2-kube-api-access-vlbqc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.727360 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6657k" event={"ID":"86c42eb0-b98f-4f6a-9107-0ece0511ccb2","Type":"ContainerDied","Data":"c3ee4395dcf44277e48702a831532113aa3f4ce8d2a0a309a1367188a61f025b"} Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.727396 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3ee4395dcf44277e48702a831532113aa3f4ce8d2a0a309a1367188a61f025b" Sep 29 10:39:48 crc kubenswrapper[4727]: I0929 10:39:48.727450 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6657k" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.086261 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.166881 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc\") pod \"39f8267e-103f-4752-847e-d82db8663bab\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.167022 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config\") pod \"39f8267e-103f-4752-847e-d82db8663bab\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.167083 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-622ws\" (UniqueName: \"kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws\") pod \"39f8267e-103f-4752-847e-d82db8663bab\" (UID: \"39f8267e-103f-4752-847e-d82db8663bab\") " Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.176813 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws" (OuterVolumeSpecName: "kube-api-access-622ws") pod "39f8267e-103f-4752-847e-d82db8663bab" (UID: "39f8267e-103f-4752-847e-d82db8663bab"). InnerVolumeSpecName "kube-api-access-622ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.207291 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config" (OuterVolumeSpecName: "config") pod "39f8267e-103f-4752-847e-d82db8663bab" (UID: "39f8267e-103f-4752-847e-d82db8663bab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.209879 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39f8267e-103f-4752-847e-d82db8663bab" (UID: "39f8267e-103f-4752-847e-d82db8663bab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.268537 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-622ws\" (UniqueName: \"kubernetes.io/projected/39f8267e-103f-4752-847e-d82db8663bab-kube-api-access-622ws\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.268577 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.268586 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f8267e-103f-4752-847e-d82db8663bab-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.284038 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.333930 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ac9728f6-fdb6-47ba-a17c-4220301c2a88" containerName="galera" probeResult="failure" output=< Sep 29 10:39:49 crc kubenswrapper[4727]: wsrep_local_state_comment (Joined) differs from Synced Sep 29 10:39:49 crc kubenswrapper[4727]: > Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.740113 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" event={"ID":"39f8267e-103f-4752-847e-d82db8663bab","Type":"ContainerDied","Data":"887e68baacb6ae198440b70186b61ade9d63b69381af568dc810d871b5d7ac3f"} Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.740156 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-524xh" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.740179 4727 scope.go:117] "RemoveContainer" containerID="9ef50efa179702acd5bc874a74c3f042103c767ca1c0edd0b0039a4c4dc646aa" Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.742268 4727 generic.go:334] "Generic (PLEG): container finished" podID="0b6fcc5d-2000-448c-9749-0318107fd488" containerID="ad0359a69955155ef944dce607d0eca027d3283393b5e65d8c13989d73036e8e" exitCode=0 Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.742372 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9kxvj" event={"ID":"0b6fcc5d-2000-448c-9749-0318107fd488","Type":"ContainerDied","Data":"ad0359a69955155ef944dce607d0eca027d3283393b5e65d8c13989d73036e8e"} Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.744625 4727 generic.go:334] "Generic (PLEG): container finished" podID="5b7d9089-b633-4493-8d9a-1a98f8febda0" containerID="c0080b1618a82670df46877e015798cc5d03f7ae3f3ca6664d8acdf2cf193ca4" exitCode=0 Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.744654 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nh7m5" event={"ID":"5b7d9089-b633-4493-8d9a-1a98f8febda0","Type":"ContainerDied","Data":"c0080b1618a82670df46877e015798cc5d03f7ae3f3ca6664d8acdf2cf193ca4"} Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.815307 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:39:49 crc kubenswrapper[4727]: I0929 10:39:49.822926 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-524xh"] Sep 29 10:39:50 crc kubenswrapper[4727]: I0929 10:39:50.613613 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x86c4" podUID="36e70287-3b92-41e6-a056-fb29d1e03772" containerName="ovn-controller" probeResult="failure" output=< Sep 29 10:39:50 crc kubenswrapper[4727]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 10:39:50 crc kubenswrapper[4727]: > Sep 29 10:39:50 crc kubenswrapper[4727]: I0929 10:39:50.753322 4727 generic.go:334] "Generic (PLEG): container finished" podID="667cfa10-bc94-4788-af5d-296745385383" containerID="d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864" exitCode=0 Sep 29 10:39:50 crc kubenswrapper[4727]: I0929 10:39:50.753402 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerDied","Data":"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864"} Sep 29 10:39:50 crc kubenswrapper[4727]: I0929 10:39:50.755670 4727 generic.go:334] "Generic (PLEG): container finished" podID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerID="1f730413d55de9f8401a7955c284e7d886a5fe384f645986bcdcbe43612ec7b6" exitCode=0 Sep 29 10:39:50 crc kubenswrapper[4727]: I0929 10:39:50.755766 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerDied","Data":"1f730413d55de9f8401a7955c284e7d886a5fe384f645986bcdcbe43612ec7b6"} Sep 29 10:39:51 crc kubenswrapper[4727]: I0929 10:39:51.132974 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39f8267e-103f-4752-847e-d82db8663bab" path="/var/lib/kubelet/pods/39f8267e-103f-4752-847e-d82db8663bab/volumes" Sep 29 10:39:53 crc kubenswrapper[4727]: I0929 10:39:53.242630 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:39:53 crc kubenswrapper[4727]: E0929 10:39:53.242963 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:39:53 crc kubenswrapper[4727]: E0929 10:39:53.243016 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:39:53 crc kubenswrapper[4727]: E0929 10:39:53.243833 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:40:09.243811683 +0000 UTC m=+1079.417125115 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:39:53 crc kubenswrapper[4727]: I0929 10:39:53.975466 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.621580 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x86c4" podUID="36e70287-3b92-41e6-a056-fb29d1e03772" containerName="ovn-controller" probeResult="failure" output=< Sep 29 10:39:55 crc kubenswrapper[4727]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 10:39:55 crc kubenswrapper[4727]: > Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.627101 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.628632 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fhhrr" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.839653 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x86c4-config-m9hbp"] Sep 29 10:39:55 crc kubenswrapper[4727]: E0929 10:39:55.840075 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="dnsmasq-dns" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840095 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="dnsmasq-dns" Sep 29 10:39:55 crc kubenswrapper[4727]: E0929 10:39:55.840118 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c42eb0-b98f-4f6a-9107-0ece0511ccb2" containerName="mariadb-database-create" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840127 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c42eb0-b98f-4f6a-9107-0ece0511ccb2" containerName="mariadb-database-create" Sep 29 10:39:55 crc kubenswrapper[4727]: E0929 10:39:55.840146 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="init" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840153 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="init" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840321 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="86c42eb0-b98f-4f6a-9107-0ece0511ccb2" containerName="mariadb-database-create" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840333 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="39f8267e-103f-4752-847e-d82db8663bab" containerName="dnsmasq-dns" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.840908 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.843210 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 29 10:39:55 crc kubenswrapper[4727]: I0929 10:39:55.854840 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4-config-m9hbp"] Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.040596 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.040933 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.041148 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.041173 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp5kq\" (UniqueName: \"kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.041422 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.041457 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143249 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143306 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143327 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143370 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp5kq\" (UniqueName: \"kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143464 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143849 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.143908 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.144321 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.144429 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.146219 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.168900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp5kq\" (UniqueName: \"kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq\") pod \"ovn-controller-x86c4-config-m9hbp\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:56 crc kubenswrapper[4727]: I0929 10:39:56.460241 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.572163 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f1a8-account-create-bbh8m"] Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.574754 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.578243 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.579844 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f1a8-account-create-bbh8m"] Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.732867 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2grh2\" (UniqueName: \"kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2\") pod \"glance-f1a8-account-create-bbh8m\" (UID: \"9fd310b2-5548-4eb1-8563-aeb4d7b81f74\") " pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.834147 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2grh2\" (UniqueName: \"kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2\") pod \"glance-f1a8-account-create-bbh8m\" (UID: \"9fd310b2-5548-4eb1-8563-aeb4d7b81f74\") " pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.857482 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2grh2\" (UniqueName: \"kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2\") pod \"glance-f1a8-account-create-bbh8m\" (UID: \"9fd310b2-5548-4eb1-8563-aeb4d7b81f74\") " pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:39:59 crc kubenswrapper[4727]: I0929 10:39:59.897192 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:40:00 crc kubenswrapper[4727]: I0929 10:40:00.621406 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x86c4" podUID="36e70287-3b92-41e6-a056-fb29d1e03772" containerName="ovn-controller" probeResult="failure" output=< Sep 29 10:40:00 crc kubenswrapper[4727]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 10:40:00 crc kubenswrapper[4727]: > Sep 29 10:40:03 crc kubenswrapper[4727]: I0929 10:40:03.792619 4727 scope.go:117] "RemoveContainer" containerID="df8ff4fabf46f6ff5add17572900fe3032dc45eddb28f87fe5d681596948309c" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.227909 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nh7m5" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.244132 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9kxvj" Sep 29 10:40:05 crc kubenswrapper[4727]: E0929 10:40:05.304443 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified" Sep 29 10:40:05 crc kubenswrapper[4727]: E0929 10:40:05.304663 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:swift-ring-rebalance,Image:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,Command:[/usr/local/bin/swift-ring-tool all],Args:[],WorkingDir:/etc/swift,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CM_NAME,Value:swift-ring-files,ValueFrom:nil,},EnvVar{Name:NAMESPACE,Value:openstack,ValueFrom:nil,},EnvVar{Name:OWNER_APIVERSION,Value:swift.openstack.org/v1beta1,ValueFrom:nil,},EnvVar{Name:OWNER_KIND,Value:SwiftRing,ValueFrom:nil,},EnvVar{Name:OWNER_NAME,Value:swift-ring,ValueFrom:nil,},EnvVar{Name:OWNER_UID,Value:7264abca-969f-41d8-9b09-fab46daccbbc,ValueFrom:nil,},EnvVar{Name:SWIFT_MIN_PART_HOURS,Value:1,ValueFrom:nil,},EnvVar{Name:SWIFT_PART_POWER,Value:10,ValueFrom:nil,},EnvVar{Name:SWIFT_REPLICAS,Value:1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/swift-ring-tool,SubPath:swift-ring-tool,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:swiftconf,ReadOnly:true,MountPath:/etc/swift/swift.conf,SubPath:swift.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ring-data-devices,ReadOnly:true,MountPath:/var/lib/config-data/ring-devices,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dispersionconf,ReadOnly:true,MountPath:/etc/swift/dispersion.conf,SubPath:dispersion.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6sn76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-ring-rebalance-6jpg2_openstack(83f5e1ea-5026-48dc-a5a7-d6d841b45376): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:40:05 crc kubenswrapper[4727]: E0929 10:40:05.305922 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/swift-ring-rebalance-6jpg2" podUID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.332311 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9hjb\" (UniqueName: \"kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb\") pod \"5b7d9089-b633-4493-8d9a-1a98f8febda0\" (UID: \"5b7d9089-b633-4493-8d9a-1a98f8febda0\") " Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.332938 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kttm\" (UniqueName: \"kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm\") pod \"0b6fcc5d-2000-448c-9749-0318107fd488\" (UID: \"0b6fcc5d-2000-448c-9749-0318107fd488\") " Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.337719 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb" (OuterVolumeSpecName: "kube-api-access-q9hjb") pod "5b7d9089-b633-4493-8d9a-1a98f8febda0" (UID: "5b7d9089-b633-4493-8d9a-1a98f8febda0"). InnerVolumeSpecName "kube-api-access-q9hjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.337758 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm" (OuterVolumeSpecName: "kube-api-access-5kttm") pod "0b6fcc5d-2000-448c-9749-0318107fd488" (UID: "0b6fcc5d-2000-448c-9749-0318107fd488"). InnerVolumeSpecName "kube-api-access-5kttm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.435033 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kttm\" (UniqueName: \"kubernetes.io/projected/0b6fcc5d-2000-448c-9749-0318107fd488-kube-api-access-5kttm\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.435071 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9hjb\" (UniqueName: \"kubernetes.io/projected/5b7d9089-b633-4493-8d9a-1a98f8febda0-kube-api-access-q9hjb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.654969 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x86c4" podUID="36e70287-3b92-41e6-a056-fb29d1e03772" containerName="ovn-controller" probeResult="failure" output=< Sep 29 10:40:05 crc kubenswrapper[4727]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 29 10:40:05 crc kubenswrapper[4727]: > Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.742008 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f1a8-account-create-bbh8m"] Sep 29 10:40:05 crc kubenswrapper[4727]: W0929 10:40:05.751499 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fd310b2_5548_4eb1_8563_aeb4d7b81f74.slice/crio-409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f WatchSource:0}: Error finding container 409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f: Status 404 returned error can't find the container with id 409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.864919 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerStarted","Data":"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.865173 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.866512 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f1a8-account-create-bbh8m" event={"ID":"9fd310b2-5548-4eb1-8563-aeb4d7b81f74","Type":"ContainerStarted","Data":"409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.875269 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerStarted","Data":"e3f4c3abce76e3e7604247f7d881163b52102d07e2ccfa590484a9575c282bf0"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.876122 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.879919 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2db0e30f-219c-4539-8f02-3f64f9f23a50","Type":"ContainerStarted","Data":"a95ea20dfde84439f3196c9dee12fe0a4ce78c9447a124ccb57143b3627b6171"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.882967 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9kxvj" event={"ID":"0b6fcc5d-2000-448c-9749-0318107fd488","Type":"ContainerDied","Data":"85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.883825 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85afc552c8006d0b95d25bc5fce33d06ec6d873bbf894751b3d47f41db8140b0" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.883371 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9kxvj" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.887523 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nh7m5" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.887499 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nh7m5" event={"ID":"5b7d9089-b633-4493-8d9a-1a98f8febda0","Type":"ContainerDied","Data":"0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15"} Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.887714 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5f1b904c9a7417f4ae0c1217337fd13490a12dcd24a75b552702d517829b15" Sep 29 10:40:05 crc kubenswrapper[4727]: E0929 10:40:05.889064 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified\\\"\"" pod="openstack/swift-ring-rebalance-6jpg2" podUID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.913455 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4-config-m9hbp"] Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.914819 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=59.472316702 podStartE2EDuration="1m6.91479978s" podCreationTimestamp="2025-09-29 10:38:59 +0000 UTC" firstStartedPulling="2025-09-29 10:39:11.913378464 +0000 UTC m=+1022.086691826" lastFinishedPulling="2025-09-29 10:39:19.355861542 +0000 UTC m=+1029.529174904" observedRunningTime="2025-09-29 10:40:05.905535413 +0000 UTC m=+1076.078848795" watchObservedRunningTime="2025-09-29 10:40:05.91479978 +0000 UTC m=+1076.088113162" Sep 29 10:40:05 crc kubenswrapper[4727]: W0929 10:40:05.915031 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0845949a_e6b9_47b8_ba8a_8b11d991ecba.slice/crio-75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0 WatchSource:0}: Error finding container 75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0: Status 404 returned error can't find the container with id 75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0 Sep 29 10:40:05 crc kubenswrapper[4727]: I0929 10:40:05.963359 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=59.618699677 podStartE2EDuration="1m6.963327675s" podCreationTimestamp="2025-09-29 10:38:59 +0000 UTC" firstStartedPulling="2025-09-29 10:39:12.494468562 +0000 UTC m=+1022.667781924" lastFinishedPulling="2025-09-29 10:39:19.83909656 +0000 UTC m=+1030.012409922" observedRunningTime="2025-09-29 10:40:05.959756073 +0000 UTC m=+1076.133069435" watchObservedRunningTime="2025-09-29 10:40:05.963327675 +0000 UTC m=+1076.136641037" Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.895623 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2db0e30f-219c-4539-8f02-3f64f9f23a50","Type":"ContainerStarted","Data":"ec64ead9f01b8c1f530c5f0c55502c730e462a1a4ecb6e16ed4cd7aaa7122041"} Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.895996 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.897538 4727 generic.go:334] "Generic (PLEG): container finished" podID="0845949a-e6b9-47b8-ba8a-8b11d991ecba" containerID="1fc13bdbedea3ddbc2b9e60abc5c887146237a6a2a5d7e340283a3d33ad137cd" exitCode=0 Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.897588 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-m9hbp" event={"ID":"0845949a-e6b9-47b8-ba8a-8b11d991ecba","Type":"ContainerDied","Data":"1fc13bdbedea3ddbc2b9e60abc5c887146237a6a2a5d7e340283a3d33ad137cd"} Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.897642 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-m9hbp" event={"ID":"0845949a-e6b9-47b8-ba8a-8b11d991ecba","Type":"ContainerStarted","Data":"75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0"} Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.899184 4727 generic.go:334] "Generic (PLEG): container finished" podID="9fd310b2-5548-4eb1-8563-aeb4d7b81f74" containerID="9c209de6c8dc0aa4cf8fe51b198a321f348fffa9ef67f1a5703e153bf05f5d47" exitCode=0 Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.899278 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f1a8-account-create-bbh8m" event={"ID":"9fd310b2-5548-4eb1-8563-aeb4d7b81f74","Type":"ContainerDied","Data":"9c209de6c8dc0aa4cf8fe51b198a321f348fffa9ef67f1a5703e153bf05f5d47"} Sep 29 10:40:06 crc kubenswrapper[4727]: I0929 10:40:06.933978 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.297272881 podStartE2EDuration="29.933956087s" podCreationTimestamp="2025-09-29 10:39:37 +0000 UTC" firstStartedPulling="2025-09-29 10:39:38.668142357 +0000 UTC m=+1048.841455719" lastFinishedPulling="2025-09-29 10:40:05.304825563 +0000 UTC m=+1075.478138925" observedRunningTime="2025-09-29 10:40:06.928388325 +0000 UTC m=+1077.101701687" watchObservedRunningTime="2025-09-29 10:40:06.933956087 +0000 UTC m=+1077.107269459" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.274076 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.281550 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.397897 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2grh2\" (UniqueName: \"kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2\") pod \"9fd310b2-5548-4eb1-8563-aeb4d7b81f74\" (UID: \"9fd310b2-5548-4eb1-8563-aeb4d7b81f74\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.397996 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp5kq\" (UniqueName: \"kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.398036 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.398085 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.398177 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.398235 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.398268 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts\") pod \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\" (UID: \"0845949a-e6b9-47b8-ba8a-8b11d991ecba\") " Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.399769 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts" (OuterVolumeSpecName: "scripts") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.399807 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.399825 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.399840 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run" (OuterVolumeSpecName: "var-run") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.400052 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.404280 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2" (OuterVolumeSpecName: "kube-api-access-2grh2") pod "9fd310b2-5548-4eb1-8563-aeb4d7b81f74" (UID: "9fd310b2-5548-4eb1-8563-aeb4d7b81f74"). InnerVolumeSpecName "kube-api-access-2grh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.412725 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq" (OuterVolumeSpecName: "kube-api-access-jp5kq") pod "0845949a-e6b9-47b8-ba8a-8b11d991ecba" (UID: "0845949a-e6b9-47b8-ba8a-8b11d991ecba"). InnerVolumeSpecName "kube-api-access-jp5kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499791 4727 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499840 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0845949a-e6b9-47b8-ba8a-8b11d991ecba-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499854 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2grh2\" (UniqueName: \"kubernetes.io/projected/9fd310b2-5548-4eb1-8563-aeb4d7b81f74-kube-api-access-2grh2\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499869 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp5kq\" (UniqueName: \"kubernetes.io/projected/0845949a-e6b9-47b8-ba8a-8b11d991ecba-kube-api-access-jp5kq\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499878 4727 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499885 4727 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.499894 4727 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0845949a-e6b9-47b8-ba8a-8b11d991ecba-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.915546 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f1a8-account-create-bbh8m" event={"ID":"9fd310b2-5548-4eb1-8563-aeb4d7b81f74","Type":"ContainerDied","Data":"409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f"} Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.915585 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f1a8-account-create-bbh8m" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.915601 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="409f98deb22ba2a1868247bc665a550b2cb8f7831851e9a7faf5a22a833b302f" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.917037 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-m9hbp" event={"ID":"0845949a-e6b9-47b8-ba8a-8b11d991ecba","Type":"ContainerDied","Data":"75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0"} Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.917077 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75407801521da224d8417b63bc377397d21e2c458e5b8e623beddb09d19b58b0" Sep 29 10:40:08 crc kubenswrapper[4727]: I0929 10:40:08.917099 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-m9hbp" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.313093 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.313356 4727 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.313393 4727 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.313463 4727 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift podName:7a6a9dee-ca84-4c2d-abc9-613bccd90764 nodeName:}" failed. No retries permitted until 2025-09-29 10:40:41.313437989 +0000 UTC m=+1111.486751351 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift") pod "swift-storage-0" (UID: "7a6a9dee-ca84-4c2d-abc9-613bccd90764") : configmap "swift-ring-files" not found Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.398811 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x86c4-config-m9hbp"] Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.408040 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x86c4-config-m9hbp"] Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.581977 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x86c4-config-7vp2g"] Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.582584 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd310b2-5548-4eb1-8563-aeb4d7b81f74" containerName="mariadb-account-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582597 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd310b2-5548-4eb1-8563-aeb4d7b81f74" containerName="mariadb-account-create" Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.582610 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0845949a-e6b9-47b8-ba8a-8b11d991ecba" containerName="ovn-config" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582617 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0845949a-e6b9-47b8-ba8a-8b11d991ecba" containerName="ovn-config" Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.582626 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6fcc5d-2000-448c-9749-0318107fd488" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582632 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6fcc5d-2000-448c-9749-0318107fd488" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: E0929 10:40:09.582649 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b7d9089-b633-4493-8d9a-1a98f8febda0" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582655 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b7d9089-b633-4493-8d9a-1a98f8febda0" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582813 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b7d9089-b633-4493-8d9a-1a98f8febda0" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582829 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd310b2-5548-4eb1-8563-aeb4d7b81f74" containerName="mariadb-account-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582841 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0845949a-e6b9-47b8-ba8a-8b11d991ecba" containerName="ovn-config" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.582850 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6fcc5d-2000-448c-9749-0318107fd488" containerName="mariadb-database-create" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.584475 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.586450 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.600208 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4-config-7vp2g"] Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722652 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722730 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722772 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722822 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7xqt\" (UniqueName: \"kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722844 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.722996 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.821743 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-vjhq6"] Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.822932 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824074 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824136 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824172 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7xqt\" (UniqueName: \"kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824191 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824218 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824272 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824519 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vlvqq" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.824652 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.825075 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.825306 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.825448 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.826144 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.828085 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.837287 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vjhq6"] Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.846565 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7xqt\" (UniqueName: \"kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt\") pod \"ovn-controller-x86c4-config-7vp2g\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.904059 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.925412 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.925543 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.925613 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:09 crc kubenswrapper[4727]: I0929 10:40:09.925672 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkp4n\" (UniqueName: \"kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.029922 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.030517 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.030568 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.030606 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkp4n\" (UniqueName: \"kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.038763 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.038863 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.045831 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.066552 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkp4n\" (UniqueName: \"kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n\") pod \"glance-db-sync-vjhq6\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.140697 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.448999 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x86c4-config-7vp2g"] Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.623920 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-x86c4" Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.828292 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vjhq6"] Sep 29 10:40:10 crc kubenswrapper[4727]: W0929 10:40:10.836765 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ed8645d_bd8d_46b7_8ccf_74ea2ed30e22.slice/crio-ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742 WatchSource:0}: Error finding container ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742: Status 404 returned error can't find the container with id ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742 Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.937285 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vjhq6" event={"ID":"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22","Type":"ContainerStarted","Data":"ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742"} Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.939246 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-7vp2g" event={"ID":"6b53c60c-15b7-43dd-b2a7-645321d62431","Type":"ContainerStarted","Data":"3bcc578aa5ae896207b924981b8d78582d68cbd8e56462debf537d5dcad10772"} Sep 29 10:40:10 crc kubenswrapper[4727]: I0929 10:40:10.939277 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-7vp2g" event={"ID":"6b53c60c-15b7-43dd-b2a7-645321d62431","Type":"ContainerStarted","Data":"72d10bfcb675691cee6974be6ab1a0688ce10f16525875db0f0569dcf3758430"} Sep 29 10:40:11 crc kubenswrapper[4727]: I0929 10:40:11.138590 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0845949a-e6b9-47b8-ba8a-8b11d991ecba" path="/var/lib/kubelet/pods/0845949a-e6b9-47b8-ba8a-8b11d991ecba/volumes" Sep 29 10:40:11 crc kubenswrapper[4727]: I0929 10:40:11.162800 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-x86c4-config-7vp2g" podStartSLOduration=2.162772049 podStartE2EDuration="2.162772049s" podCreationTimestamp="2025-09-29 10:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:10.964817474 +0000 UTC m=+1081.138130836" watchObservedRunningTime="2025-09-29 10:40:11.162772049 +0000 UTC m=+1081.336085411" Sep 29 10:40:11 crc kubenswrapper[4727]: I0929 10:40:11.952399 4727 generic.go:334] "Generic (PLEG): container finished" podID="6b53c60c-15b7-43dd-b2a7-645321d62431" containerID="3bcc578aa5ae896207b924981b8d78582d68cbd8e56462debf537d5dcad10772" exitCode=0 Sep 29 10:40:11 crc kubenswrapper[4727]: I0929 10:40:11.952462 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x86c4-config-7vp2g" event={"ID":"6b53c60c-15b7-43dd-b2a7-645321d62431","Type":"ContainerDied","Data":"3bcc578aa5ae896207b924981b8d78582d68cbd8e56462debf537d5dcad10772"} Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.400023 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494596 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7xqt\" (UniqueName: \"kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494663 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494690 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494790 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494827 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494851 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494875 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run\") pod \"6b53c60c-15b7-43dd-b2a7-645321d62431\" (UID: \"6b53c60c-15b7-43dd-b2a7-645321d62431\") " Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494914 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run" (OuterVolumeSpecName: "var-run") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.494974 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495600 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495746 4727 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495771 4727 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495784 4727 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495798 4727 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b53c60c-15b7-43dd-b2a7-645321d62431-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.495818 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts" (OuterVolumeSpecName: "scripts") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.504779 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt" (OuterVolumeSpecName: "kube-api-access-j7xqt") pod "6b53c60c-15b7-43dd-b2a7-645321d62431" (UID: "6b53c60c-15b7-43dd-b2a7-645321d62431"). InnerVolumeSpecName "kube-api-access-j7xqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.568081 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x86c4-config-7vp2g"] Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.578126 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x86c4-config-7vp2g"] Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.597580 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b53c60c-15b7-43dd-b2a7-645321d62431-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.597614 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7xqt\" (UniqueName: \"kubernetes.io/projected/6b53c60c-15b7-43dd-b2a7-645321d62431-kube-api-access-j7xqt\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.961366 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-81c7-account-create-hqw9d"] Sep 29 10:40:13 crc kubenswrapper[4727]: E0929 10:40:13.961809 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b53c60c-15b7-43dd-b2a7-645321d62431" containerName="ovn-config" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.961831 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b53c60c-15b7-43dd-b2a7-645321d62431" containerName="ovn-config" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.962029 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b53c60c-15b7-43dd-b2a7-645321d62431" containerName="ovn-config" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.962684 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.965855 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.972062 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81c7-account-create-hqw9d"] Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.982982 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72d10bfcb675691cee6974be6ab1a0688ce10f16525875db0f0569dcf3758430" Sep 29 10:40:13 crc kubenswrapper[4727]: I0929 10:40:13.983039 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x86c4-config-7vp2g" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.105917 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9k9w\" (UniqueName: \"kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w\") pod \"keystone-81c7-account-create-hqw9d\" (UID: \"3425b874-dcfd-47e3-bf52-4510b6ab77bb\") " pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.209520 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9k9w\" (UniqueName: \"kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w\") pod \"keystone-81c7-account-create-hqw9d\" (UID: \"3425b874-dcfd-47e3-bf52-4510b6ab77bb\") " pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.227412 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9k9w\" (UniqueName: \"kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w\") pod \"keystone-81c7-account-create-hqw9d\" (UID: \"3425b874-dcfd-47e3-bf52-4510b6ab77bb\") " pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.266679 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75ca-account-create-8gcdc"] Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.268110 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.270304 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.280229 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75ca-account-create-8gcdc"] Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.290220 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.412849 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77579\" (UniqueName: \"kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579\") pod \"placement-75ca-account-create-8gcdc\" (UID: \"3b4d58c3-da91-44ca-af8b-d542b285332a\") " pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.515263 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77579\" (UniqueName: \"kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579\") pod \"placement-75ca-account-create-8gcdc\" (UID: \"3b4d58c3-da91-44ca-af8b-d542b285332a\") " pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.533947 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77579\" (UniqueName: \"kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579\") pod \"placement-75ca-account-create-8gcdc\" (UID: \"3b4d58c3-da91-44ca-af8b-d542b285332a\") " pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.594033 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.761961 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81c7-account-create-hqw9d"] Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.995983 4727 generic.go:334] "Generic (PLEG): container finished" podID="3425b874-dcfd-47e3-bf52-4510b6ab77bb" containerID="c803ad8f87069d1b6a8b4b56f280c32fd59952bdab95195e3cfd4150cdf6ce00" exitCode=0 Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.996057 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81c7-account-create-hqw9d" event={"ID":"3425b874-dcfd-47e3-bf52-4510b6ab77bb","Type":"ContainerDied","Data":"c803ad8f87069d1b6a8b4b56f280c32fd59952bdab95195e3cfd4150cdf6ce00"} Sep 29 10:40:14 crc kubenswrapper[4727]: I0929 10:40:14.996114 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81c7-account-create-hqw9d" event={"ID":"3425b874-dcfd-47e3-bf52-4510b6ab77bb","Type":"ContainerStarted","Data":"f2ac9258843fe395a0ac033cadfae60124d27075589e2dd47fbf79b9ce119e4f"} Sep 29 10:40:15 crc kubenswrapper[4727]: I0929 10:40:15.021420 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75ca-account-create-8gcdc"] Sep 29 10:40:15 crc kubenswrapper[4727]: W0929 10:40:15.030832 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b4d58c3_da91_44ca_af8b_d542b285332a.slice/crio-d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f WatchSource:0}: Error finding container d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f: Status 404 returned error can't find the container with id d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f Sep 29 10:40:15 crc kubenswrapper[4727]: I0929 10:40:15.119293 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b53c60c-15b7-43dd-b2a7-645321d62431" path="/var/lib/kubelet/pods/6b53c60c-15b7-43dd-b2a7-645321d62431/volumes" Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.006224 4727 generic.go:334] "Generic (PLEG): container finished" podID="3b4d58c3-da91-44ca-af8b-d542b285332a" containerID="08b8445010b3cd7b845a00a34267d69e84e4830688094a879c4a39847be1cde4" exitCode=0 Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.006316 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75ca-account-create-8gcdc" event={"ID":"3b4d58c3-da91-44ca-af8b-d542b285332a","Type":"ContainerDied","Data":"08b8445010b3cd7b845a00a34267d69e84e4830688094a879c4a39847be1cde4"} Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.006671 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75ca-account-create-8gcdc" event={"ID":"3b4d58c3-da91-44ca-af8b-d542b285332a","Type":"ContainerStarted","Data":"d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f"} Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.337500 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.455305 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9k9w\" (UniqueName: \"kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w\") pod \"3425b874-dcfd-47e3-bf52-4510b6ab77bb\" (UID: \"3425b874-dcfd-47e3-bf52-4510b6ab77bb\") " Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.463704 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w" (OuterVolumeSpecName: "kube-api-access-l9k9w") pod "3425b874-dcfd-47e3-bf52-4510b6ab77bb" (UID: "3425b874-dcfd-47e3-bf52-4510b6ab77bb"). InnerVolumeSpecName "kube-api-access-l9k9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:16 crc kubenswrapper[4727]: I0929 10:40:16.558315 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9k9w\" (UniqueName: \"kubernetes.io/projected/3425b874-dcfd-47e3-bf52-4510b6ab77bb-kube-api-access-l9k9w\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:17 crc kubenswrapper[4727]: I0929 10:40:17.016041 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81c7-account-create-hqw9d" Sep 29 10:40:17 crc kubenswrapper[4727]: I0929 10:40:17.016040 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81c7-account-create-hqw9d" event={"ID":"3425b874-dcfd-47e3-bf52-4510b6ab77bb","Type":"ContainerDied","Data":"f2ac9258843fe395a0ac033cadfae60124d27075589e2dd47fbf79b9ce119e4f"} Sep 29 10:40:17 crc kubenswrapper[4727]: I0929 10:40:17.016521 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2ac9258843fe395a0ac033cadfae60124d27075589e2dd47fbf79b9ce119e4f" Sep 29 10:40:18 crc kubenswrapper[4727]: I0929 10:40:18.243217 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 29 10:40:20 crc kubenswrapper[4727]: I0929 10:40:20.902556 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:40:21 crc kubenswrapper[4727]: I0929 10:40:21.268547 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.634526 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-kmtq9"] Sep 29 10:40:22 crc kubenswrapper[4727]: E0929 10:40:22.634860 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3425b874-dcfd-47e3-bf52-4510b6ab77bb" containerName="mariadb-account-create" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.634872 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="3425b874-dcfd-47e3-bf52-4510b6ab77bb" containerName="mariadb-account-create" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.635039 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="3425b874-dcfd-47e3-bf52-4510b6ab77bb" containerName="mariadb-account-create" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.635633 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.657921 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kmtq9"] Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.680438 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btlnq\" (UniqueName: \"kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq\") pod \"barbican-db-create-kmtq9\" (UID: \"68b33a86-c4c4-469f-8168-0c9a84bce521\") " pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.782166 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btlnq\" (UniqueName: \"kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq\") pod \"barbican-db-create-kmtq9\" (UID: \"68b33a86-c4c4-469f-8168-0c9a84bce521\") " pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.809148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btlnq\" (UniqueName: \"kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq\") pod \"barbican-db-create-kmtq9\" (UID: \"68b33a86-c4c4-469f-8168-0c9a84bce521\") " pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.840893 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bnj9q"] Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.842240 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.861494 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bnj9q"] Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.884288 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88vzg\" (UniqueName: \"kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg\") pod \"cinder-db-create-bnj9q\" (UID: \"78290c73-5265-43d0-825d-5b9e3164a754\") " pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.924459 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6bchk"] Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.925404 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.940743 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6bchk"] Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.952743 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.985433 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88vzg\" (UniqueName: \"kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg\") pod \"cinder-db-create-bnj9q\" (UID: \"78290c73-5265-43d0-825d-5b9e3164a754\") " pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:22 crc kubenswrapper[4727]: I0929 10:40:22.985573 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jfjt\" (UniqueName: \"kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt\") pod \"neutron-db-create-6bchk\" (UID: \"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d\") " pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.000654 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88vzg\" (UniqueName: \"kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg\") pod \"cinder-db-create-bnj9q\" (UID: \"78290c73-5265-43d0-825d-5b9e3164a754\") " pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.079437 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zl5lc"] Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.080783 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.085310 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.085550 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2g6jb" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.085578 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.085707 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.086942 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jfjt\" (UniqueName: \"kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt\") pod \"neutron-db-create-6bchk\" (UID: \"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d\") " pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.100540 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zl5lc"] Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.107880 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jfjt\" (UniqueName: \"kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt\") pod \"neutron-db-create-6bchk\" (UID: \"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d\") " pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.161380 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.188577 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdgrm\" (UniqueName: \"kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.188624 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.188734 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.245288 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.289738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.289860 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdgrm\" (UniqueName: \"kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.289896 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.293825 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.294687 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.310816 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdgrm\" (UniqueName: \"kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm\") pod \"keystone-db-sync-zl5lc\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.408035 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.675427 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.798427 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77579\" (UniqueName: \"kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579\") pod \"3b4d58c3-da91-44ca-af8b-d542b285332a\" (UID: \"3b4d58c3-da91-44ca-af8b-d542b285332a\") " Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.804392 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579" (OuterVolumeSpecName: "kube-api-access-77579") pod "3b4d58c3-da91-44ca-af8b-d542b285332a" (UID: "3b4d58c3-da91-44ca-af8b-d542b285332a"). InnerVolumeSpecName "kube-api-access-77579". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:23 crc kubenswrapper[4727]: I0929 10:40:23.901166 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77579\" (UniqueName: \"kubernetes.io/projected/3b4d58c3-da91-44ca-af8b-d542b285332a-kube-api-access-77579\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.108014 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75ca-account-create-8gcdc" event={"ID":"3b4d58c3-da91-44ca-af8b-d542b285332a","Type":"ContainerDied","Data":"d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f"} Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.108345 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8940f7550dd196f65fab3182af965f485a62c26cc649b44c78c5e1d7a25323f" Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.108297 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75ca-account-create-8gcdc" Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.234738 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kmtq9"] Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.268860 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6bchk"] Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.358461 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zl5lc"] Sep 29 10:40:24 crc kubenswrapper[4727]: W0929 10:40:24.362754 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda838d423_a730_48fe_b345_9e11d8f290a6.slice/crio-367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8 WatchSource:0}: Error finding container 367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8: Status 404 returned error can't find the container with id 367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8 Sep 29 10:40:24 crc kubenswrapper[4727]: I0929 10:40:24.455995 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bnj9q"] Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.119861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zl5lc" event={"ID":"a838d423-a730-48fe-b345-9e11d8f290a6","Type":"ContainerStarted","Data":"367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.122107 4727 generic.go:334] "Generic (PLEG): container finished" podID="78290c73-5265-43d0-825d-5b9e3164a754" containerID="def633393858396e91b3eaeebae4c6955672243cac3c887c2a584bb8bb4fe39f" exitCode=0 Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.122162 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bnj9q" event={"ID":"78290c73-5265-43d0-825d-5b9e3164a754","Type":"ContainerDied","Data":"def633393858396e91b3eaeebae4c6955672243cac3c887c2a584bb8bb4fe39f"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.122179 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bnj9q" event={"ID":"78290c73-5265-43d0-825d-5b9e3164a754","Type":"ContainerStarted","Data":"638ca7f208305fc6c945bca792be6454a5e49a7dc06187b855a59801634b4fad"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.124283 4727 generic.go:334] "Generic (PLEG): container finished" podID="68b33a86-c4c4-469f-8168-0c9a84bce521" containerID="71e7bcd6121589881bbcccf09df7a13ba17446dc9e858f0ebfcacd8579bcc0c9" exitCode=0 Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.124372 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kmtq9" event={"ID":"68b33a86-c4c4-469f-8168-0c9a84bce521","Type":"ContainerDied","Data":"71e7bcd6121589881bbcccf09df7a13ba17446dc9e858f0ebfcacd8579bcc0c9"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.124402 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kmtq9" event={"ID":"68b33a86-c4c4-469f-8168-0c9a84bce521","Type":"ContainerStarted","Data":"f95ac6a0e9a82e5f2fd9d86618aaa7114c64a555fbd2cb7868b681ef699fbfa4"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.128635 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6jpg2" event={"ID":"83f5e1ea-5026-48dc-a5a7-d6d841b45376","Type":"ContainerStarted","Data":"99b64a3cde9e5f383ac79c48bd904d75af53d4c1bfed89ff3fcc5a4b051b6712"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.132949 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vjhq6" event={"ID":"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22","Type":"ContainerStarted","Data":"e68db5791dc8003f4e619fea8afd301620b3a99fbace6a84e941c3f5b1d36cb0"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.135195 4727 generic.go:334] "Generic (PLEG): container finished" podID="36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" containerID="2549db53c32fed963c223d4ee19940e442f6d961fbbc8c9f381f63fd4fa65bd7" exitCode=0 Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.135328 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bchk" event={"ID":"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d","Type":"ContainerDied","Data":"2549db53c32fed963c223d4ee19940e442f6d961fbbc8c9f381f63fd4fa65bd7"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.135371 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bchk" event={"ID":"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d","Type":"ContainerStarted","Data":"fbfcbc355b9fa0b72aaafd63236608b70dc8ef11340303786a1ea9c442924ad5"} Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.160600 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-vjhq6" podStartSLOduration=3.175620773 podStartE2EDuration="16.16057739s" podCreationTimestamp="2025-09-29 10:40:09 +0000 UTC" firstStartedPulling="2025-09-29 10:40:10.83945844 +0000 UTC m=+1081.012771802" lastFinishedPulling="2025-09-29 10:40:23.824415057 +0000 UTC m=+1093.997728419" observedRunningTime="2025-09-29 10:40:25.15237756 +0000 UTC m=+1095.325690922" watchObservedRunningTime="2025-09-29 10:40:25.16057739 +0000 UTC m=+1095.333890752" Sep 29 10:40:25 crc kubenswrapper[4727]: I0929 10:40:25.170810 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6jpg2" podStartSLOduration=2.374330475 podStartE2EDuration="44.170789292s" podCreationTimestamp="2025-09-29 10:39:41 +0000 UTC" firstStartedPulling="2025-09-29 10:39:42.013051978 +0000 UTC m=+1052.186365340" lastFinishedPulling="2025-09-29 10:40:23.809510795 +0000 UTC m=+1093.982824157" observedRunningTime="2025-09-29 10:40:25.166563944 +0000 UTC m=+1095.339877326" watchObservedRunningTime="2025-09-29 10:40:25.170789292 +0000 UTC m=+1095.344102654" Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.860362 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.867256 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.878294 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.999291 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btlnq\" (UniqueName: \"kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq\") pod \"68b33a86-c4c4-469f-8168-0c9a84bce521\" (UID: \"68b33a86-c4c4-469f-8168-0c9a84bce521\") " Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.999711 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jfjt\" (UniqueName: \"kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt\") pod \"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d\" (UID: \"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d\") " Sep 29 10:40:31 crc kubenswrapper[4727]: I0929 10:40:31.999770 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88vzg\" (UniqueName: \"kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg\") pod \"78290c73-5265-43d0-825d-5b9e3164a754\" (UID: \"78290c73-5265-43d0-825d-5b9e3164a754\") " Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.002830 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq" (OuterVolumeSpecName: "kube-api-access-btlnq") pod "68b33a86-c4c4-469f-8168-0c9a84bce521" (UID: "68b33a86-c4c4-469f-8168-0c9a84bce521"). InnerVolumeSpecName "kube-api-access-btlnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.003829 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg" (OuterVolumeSpecName: "kube-api-access-88vzg") pod "78290c73-5265-43d0-825d-5b9e3164a754" (UID: "78290c73-5265-43d0-825d-5b9e3164a754"). InnerVolumeSpecName "kube-api-access-88vzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.003915 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt" (OuterVolumeSpecName: "kube-api-access-2jfjt") pod "36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" (UID: "36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d"). InnerVolumeSpecName "kube-api-access-2jfjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.101846 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jfjt\" (UniqueName: \"kubernetes.io/projected/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d-kube-api-access-2jfjt\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.101877 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88vzg\" (UniqueName: \"kubernetes.io/projected/78290c73-5265-43d0-825d-5b9e3164a754-kube-api-access-88vzg\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.101886 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btlnq\" (UniqueName: \"kubernetes.io/projected/68b33a86-c4c4-469f-8168-0c9a84bce521-kube-api-access-btlnq\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.206227 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zl5lc" event={"ID":"a838d423-a730-48fe-b345-9e11d8f290a6","Type":"ContainerStarted","Data":"e65795c49dc721454002fc44688fe81f9664c5b5dcb7d72e0384d0feca42c8e8"} Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.208577 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bnj9q" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.208568 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bnj9q" event={"ID":"78290c73-5265-43d0-825d-5b9e3164a754","Type":"ContainerDied","Data":"638ca7f208305fc6c945bca792be6454a5e49a7dc06187b855a59801634b4fad"} Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.208712 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="638ca7f208305fc6c945bca792be6454a5e49a7dc06187b855a59801634b4fad" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.209897 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kmtq9" event={"ID":"68b33a86-c4c4-469f-8168-0c9a84bce521","Type":"ContainerDied","Data":"f95ac6a0e9a82e5f2fd9d86618aaa7114c64a555fbd2cb7868b681ef699fbfa4"} Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.209999 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f95ac6a0e9a82e5f2fd9d86618aaa7114c64a555fbd2cb7868b681ef699fbfa4" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.209913 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kmtq9" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.211023 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6bchk" event={"ID":"36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d","Type":"ContainerDied","Data":"fbfcbc355b9fa0b72aaafd63236608b70dc8ef11340303786a1ea9c442924ad5"} Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.211055 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbfcbc355b9fa0b72aaafd63236608b70dc8ef11340303786a1ea9c442924ad5" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.211061 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6bchk" Sep 29 10:40:32 crc kubenswrapper[4727]: I0929 10:40:32.219551 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zl5lc" podStartSLOduration=1.710428823 podStartE2EDuration="9.219536286s" podCreationTimestamp="2025-09-29 10:40:23 +0000 UTC" firstStartedPulling="2025-09-29 10:40:24.365284193 +0000 UTC m=+1094.538597555" lastFinishedPulling="2025-09-29 10:40:31.874391656 +0000 UTC m=+1102.047705018" observedRunningTime="2025-09-29 10:40:32.217896944 +0000 UTC m=+1102.391210306" watchObservedRunningTime="2025-09-29 10:40:32.219536286 +0000 UTC m=+1102.392849648" Sep 29 10:40:33 crc kubenswrapper[4727]: I0929 10:40:33.220328 4727 generic.go:334] "Generic (PLEG): container finished" podID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" containerID="99b64a3cde9e5f383ac79c48bd904d75af53d4c1bfed89ff3fcc5a4b051b6712" exitCode=0 Sep 29 10:40:33 crc kubenswrapper[4727]: I0929 10:40:33.221360 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6jpg2" event={"ID":"83f5e1ea-5026-48dc-a5a7-d6d841b45376","Type":"ContainerDied","Data":"99b64a3cde9e5f383ac79c48bd904d75af53d4c1bfed89ff3fcc5a4b051b6712"} Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.234533 4727 generic.go:334] "Generic (PLEG): container finished" podID="7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" containerID="e68db5791dc8003f4e619fea8afd301620b3a99fbace6a84e941c3f5b1d36cb0" exitCode=0 Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.234632 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vjhq6" event={"ID":"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22","Type":"ContainerDied","Data":"e68db5791dc8003f4e619fea8afd301620b3a99fbace6a84e941c3f5b1d36cb0"} Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.570926 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753622 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753710 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753789 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753821 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753861 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753892 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.753946 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sn76\" (UniqueName: \"kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76\") pod \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\" (UID: \"83f5e1ea-5026-48dc-a5a7-d6d841b45376\") " Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.755478 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.755522 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.759175 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76" (OuterVolumeSpecName: "kube-api-access-6sn76") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "kube-api-access-6sn76". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.776489 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts" (OuterVolumeSpecName: "scripts") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.777933 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.778853 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.779925 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "83f5e1ea-5026-48dc-a5a7-d6d841b45376" (UID: "83f5e1ea-5026-48dc-a5a7-d6d841b45376"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.855956 4727 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83f5e1ea-5026-48dc-a5a7-d6d841b45376-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.855993 4727 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.856004 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.856016 4727 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83f5e1ea-5026-48dc-a5a7-d6d841b45376-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.856025 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.856034 4727 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83f5e1ea-5026-48dc-a5a7-d6d841b45376-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:34 crc kubenswrapper[4727]: I0929 10:40:34.856058 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sn76\" (UniqueName: \"kubernetes.io/projected/83f5e1ea-5026-48dc-a5a7-d6d841b45376-kube-api-access-6sn76\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.243452 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6jpg2" event={"ID":"83f5e1ea-5026-48dc-a5a7-d6d841b45376","Type":"ContainerDied","Data":"2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750"} Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.243493 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dd82c9de41cbb9b8fdc4f774e30b2fc9cd71047f636fd2cdfd4a7c7116dd750" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.243495 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6jpg2" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.617078 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.776822 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data\") pod \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.777008 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data\") pod \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.777055 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkp4n\" (UniqueName: \"kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n\") pod \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.777075 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle\") pod \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\" (UID: \"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22\") " Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.783142 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" (UID: "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.783726 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n" (OuterVolumeSpecName: "kube-api-access-pkp4n") pod "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" (UID: "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22"). InnerVolumeSpecName "kube-api-access-pkp4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.804507 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" (UID: "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.838871 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data" (OuterVolumeSpecName: "config-data") pod "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" (UID: "7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.879079 4727 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.879120 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkp4n\" (UniqueName: \"kubernetes.io/projected/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-kube-api-access-pkp4n\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.879136 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:35 crc kubenswrapper[4727]: I0929 10:40:35.879146 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.251942 4727 generic.go:334] "Generic (PLEG): container finished" podID="a838d423-a730-48fe-b345-9e11d8f290a6" containerID="e65795c49dc721454002fc44688fe81f9664c5b5dcb7d72e0384d0feca42c8e8" exitCode=0 Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.252031 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zl5lc" event={"ID":"a838d423-a730-48fe-b345-9e11d8f290a6","Type":"ContainerDied","Data":"e65795c49dc721454002fc44688fe81f9664c5b5dcb7d72e0384d0feca42c8e8"} Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.254089 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vjhq6" event={"ID":"7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22","Type":"ContainerDied","Data":"ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742"} Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.254131 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce19a7d6cc7a3cd3ba3e519973a90d45d61d2c881c4d00774ba02800fb398742" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.254168 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vjhq6" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.668297 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669290 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" containerName="swift-ring-rebalance" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669309 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" containerName="swift-ring-rebalance" Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669320 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4d58c3-da91-44ca-af8b-d542b285332a" containerName="mariadb-account-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669327 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4d58c3-da91-44ca-af8b-d542b285332a" containerName="mariadb-account-create" Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669363 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" containerName="glance-db-sync" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669372 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" containerName="glance-db-sync" Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669392 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b33a86-c4c4-469f-8168-0c9a84bce521" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669399 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b33a86-c4c4-469f-8168-0c9a84bce521" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669409 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78290c73-5265-43d0-825d-5b9e3164a754" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669417 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="78290c73-5265-43d0-825d-5b9e3164a754" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: E0929 10:40:36.669427 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669434 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669666 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669690 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f5e1ea-5026-48dc-a5a7-d6d841b45376" containerName="swift-ring-rebalance" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669701 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4d58c3-da91-44ca-af8b-d542b285332a" containerName="mariadb-account-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669715 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" containerName="glance-db-sync" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669728 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b33a86-c4c4-469f-8168-0c9a84bce521" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.669743 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="78290c73-5265-43d0-825d-5b9e3164a754" containerName="mariadb-database-create" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.672411 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.688507 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.793554 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzwlr\" (UniqueName: \"kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.793706 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.793756 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.793823 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.793890 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.896284 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzwlr\" (UniqueName: \"kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.896419 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.896470 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.896502 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.896550 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.897255 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.898144 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.898153 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.898192 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.920909 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzwlr\" (UniqueName: \"kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr\") pod \"dnsmasq-dns-5b946c75cc-jndcv\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:36 crc kubenswrapper[4727]: I0929 10:40:36.991859 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.456880 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.604062 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.712094 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdgrm\" (UniqueName: \"kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm\") pod \"a838d423-a730-48fe-b345-9e11d8f290a6\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.712780 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data\") pod \"a838d423-a730-48fe-b345-9e11d8f290a6\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.712838 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle\") pod \"a838d423-a730-48fe-b345-9e11d8f290a6\" (UID: \"a838d423-a730-48fe-b345-9e11d8f290a6\") " Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.727776 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm" (OuterVolumeSpecName: "kube-api-access-tdgrm") pod "a838d423-a730-48fe-b345-9e11d8f290a6" (UID: "a838d423-a730-48fe-b345-9e11d8f290a6"). InnerVolumeSpecName "kube-api-access-tdgrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.742005 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a838d423-a730-48fe-b345-9e11d8f290a6" (UID: "a838d423-a730-48fe-b345-9e11d8f290a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.765522 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data" (OuterVolumeSpecName: "config-data") pod "a838d423-a730-48fe-b345-9e11d8f290a6" (UID: "a838d423-a730-48fe-b345-9e11d8f290a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.814237 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.814284 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a838d423-a730-48fe-b345-9e11d8f290a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:37 crc kubenswrapper[4727]: I0929 10:40:37.814297 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdgrm\" (UniqueName: \"kubernetes.io/projected/a838d423-a730-48fe-b345-9e11d8f290a6-kube-api-access-tdgrm\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.291234 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zl5lc" event={"ID":"a838d423-a730-48fe-b345-9e11d8f290a6","Type":"ContainerDied","Data":"367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8"} Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.291312 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367d2c9951bf952834affb0c84af27a1e3c757c4f2ac9df637fe2409967996a8" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.291478 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zl5lc" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.303189 4727 generic.go:334] "Generic (PLEG): container finished" podID="be580dcd-1367-419d-be4f-0b36c258fd13" containerID="b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f" exitCode=0 Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.303277 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" event={"ID":"be580dcd-1367-419d-be4f-0b36c258fd13","Type":"ContainerDied","Data":"b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f"} Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.303315 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" event={"ID":"be580dcd-1367-419d-be4f-0b36c258fd13","Type":"ContainerStarted","Data":"41dc924481a875296642d5181a6dc44913b21b365849b2296a37a67414248c80"} Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.536992 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.577013 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mdzcd"] Sep 29 10:40:38 crc kubenswrapper[4727]: E0929 10:40:38.578881 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a838d423-a730-48fe-b345-9e11d8f290a6" containerName="keystone-db-sync" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.578919 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a838d423-a730-48fe-b345-9e11d8f290a6" containerName="keystone-db-sync" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.579373 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a838d423-a730-48fe-b345-9e11d8f290a6" containerName="keystone-db-sync" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.580472 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.607228 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.630201 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.630435 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2g6jb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.630551 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.637807 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mdzcd"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.660252 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.661825 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.676437 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734158 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734514 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8r8h\" (UniqueName: \"kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734741 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734816 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734851 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.734908 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.767963 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.769581 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.781756 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-ctvfm" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.781989 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.782552 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.782809 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.793636 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838154 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838280 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838317 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838370 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838398 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.838434 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.850137 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.850578 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.850696 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.850869 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8r8h\" (UniqueName: \"kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.850972 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.851131 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbjh7\" (UniqueName: \"kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.867432 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.868054 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.870860 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.878511 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.882462 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.883291 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8r8h\" (UniqueName: \"kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.890296 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys\") pod \"keystone-bootstrap-mdzcd\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.891054 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.891257 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.898941 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956253 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956446 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956499 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956516 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956550 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956572 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8xnt\" (UniqueName: \"kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956594 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956643 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956678 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbjh7\" (UniqueName: \"kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956703 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.956721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.960812 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.965613 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.975313 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.977358 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.979755 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cx7d2"] Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.981009 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.984598 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pcthw" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.984828 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.984974 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.994553 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbjh7\" (UniqueName: \"kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7\") pod \"dnsmasq-dns-784f69c749-t4wzb\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:38 crc kubenswrapper[4727]: I0929 10:40:38.998231 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cx7d2"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.010950 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.026394 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.028259 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.031915 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.032207 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.033134 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.040406 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vlvqq" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058273 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058353 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nphtd\" (UniqueName: \"kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058380 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058403 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058422 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058465 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058534 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058560 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058587 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058627 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058652 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8xnt\" (UniqueName: \"kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.058675 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.059611 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.059659 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.060996 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.080183 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.134133 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8xnt\" (UniqueName: \"kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt\") pod \"horizon-56b7b5f9ff-9bp57\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.156214 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.160431 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.160616 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.160744 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170018 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170275 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69bdl\" (UniqueName: \"kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170432 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170571 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170660 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nphtd\" (UniqueName: \"kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170809 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170889 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.170969 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171046 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171173 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171285 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171380 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171542 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.166678 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.172960 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.175635 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.177908 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.171675 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.182075 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tkfz\" (UniqueName: \"kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.185995 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.191127 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.194971 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.196080 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.219434 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nphtd\" (UniqueName: \"kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd\") pod \"ceilometer-0\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.228635 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284308 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284387 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284404 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69bdl\" (UniqueName: \"kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284431 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284464 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284495 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284561 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284609 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284644 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284662 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284683 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rksws\" (UniqueName: \"kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284720 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284740 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tkfz\" (UniqueName: \"kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284758 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284787 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284805 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.284847 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.293090 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.293437 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.293742 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.295123 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.305418 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.306948 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.309109 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.315537 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.317925 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.335385 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.338241 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.340385 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69bdl\" (UniqueName: \"kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.344017 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.383757 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.387513 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.387692 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rksws\" (UniqueName: \"kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.387836 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.387960 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.389150 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.389849 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.392035 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.393455 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.398857 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.400587 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tkfz\" (UniqueName: \"kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz\") pod \"placement-db-sync-cx7d2\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.402371 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.415148 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.423321 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" event={"ID":"be580dcd-1367-419d-be4f-0b36c258fd13","Type":"ContainerStarted","Data":"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70"} Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.423520 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="dnsmasq-dns" containerID="cri-o://759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70" gracePeriod=10 Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.423802 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.423975 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.461410 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.462182 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rksws\" (UniqueName: \"kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws\") pod \"horizon-64d84c44f5-8xbxm\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.475451 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.492055 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.495636 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.500146 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsf98\" (UniqueName: \"kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.500199 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.500597 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.509355 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.521568 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.552221 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cx7d2" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.565565 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" podStartSLOduration=3.565537469 podStartE2EDuration="3.565537469s" podCreationTimestamp="2025-09-29 10:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:39.480990953 +0000 UTC m=+1109.654304315" watchObservedRunningTime="2025-09-29 10:40:39.565537469 +0000 UTC m=+1109.738850831" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.599506 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.615985 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616045 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjcb5\" (UniqueName: \"kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616078 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616116 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616133 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616160 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsf98\" (UniqueName: \"kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616179 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616197 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616215 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616287 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616308 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.616328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.618964 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.620965 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.632828 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.633098 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.637137 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.651003 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsf98\" (UniqueName: \"kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98\") pod \"dnsmasq-dns-f84976bdf-sd6r4\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.718882 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.718966 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.721300 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjcb5\" (UniqueName: \"kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.721413 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.721456 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.721511 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.721550 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.722686 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.722826 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.723067 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.723655 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.725839 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.729746 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.733968 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.758065 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjcb5\" (UniqueName: \"kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.809311 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.814119 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mdzcd"] Sep 29 10:40:39 crc kubenswrapper[4727]: I0929 10:40:39.823732 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.058297 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:40 crc kubenswrapper[4727]: W0929 10:40:40.063891 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88e12d3e_6d5c_4c08_9279_cdcd18108f1e.slice/crio-87e5fe3e1478f45e19aee2b7b5e0eb85931d5a51daf6530899cee672ecb972bf WatchSource:0}: Error finding container 87e5fe3e1478f45e19aee2b7b5e0eb85931d5a51daf6530899cee672ecb972bf: Status 404 returned error can't find the container with id 87e5fe3e1478f45e19aee2b7b5e0eb85931d5a51daf6530899cee672ecb972bf Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.066619 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.278697 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cx7d2"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.303951 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.315205 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.325857 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.391071 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzwlr\" (UniqueName: \"kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr\") pod \"be580dcd-1367-419d-be4f-0b36c258fd13\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.391604 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb\") pod \"be580dcd-1367-419d-be4f-0b36c258fd13\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.391809 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb\") pod \"be580dcd-1367-419d-be4f-0b36c258fd13\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.391828 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config\") pod \"be580dcd-1367-419d-be4f-0b36c258fd13\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.391934 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc\") pod \"be580dcd-1367-419d-be4f-0b36c258fd13\" (UID: \"be580dcd-1367-419d-be4f-0b36c258fd13\") " Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.393451 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.412384 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr" (OuterVolumeSpecName: "kube-api-access-pzwlr") pod "be580dcd-1367-419d-be4f-0b36c258fd13" (UID: "be580dcd-1367-419d-be4f-0b36c258fd13"). InnerVolumeSpecName "kube-api-access-pzwlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:40 crc kubenswrapper[4727]: W0929 10:40:40.415231 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23818aef_5aaf_4848_af35_64585311f0b1.slice/crio-5292fb802ca6f9af7b2318f2c8704b8753a5ea7c4cd56f16933096cbf32c04da WatchSource:0}: Error finding container 5292fb802ca6f9af7b2318f2c8704b8753a5ea7c4cd56f16933096cbf32c04da: Status 404 returned error can't find the container with id 5292fb802ca6f9af7b2318f2c8704b8753a5ea7c4cd56f16933096cbf32c04da Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.436400 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d84c44f5-8xbxm" event={"ID":"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4","Type":"ContainerStarted","Data":"25fecaa7557d3c0dd2d3ff6f86c1820f71f0b8f30258a647fcfd9a95342f838f"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.448059 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56b7b5f9ff-9bp57" event={"ID":"88e12d3e-6d5c-4c08-9279-cdcd18108f1e","Type":"ContainerStarted","Data":"87e5fe3e1478f45e19aee2b7b5e0eb85931d5a51daf6530899cee672ecb972bf"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.450857 4727 generic.go:334] "Generic (PLEG): container finished" podID="be580dcd-1367-419d-be4f-0b36c258fd13" containerID="759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70" exitCode=0 Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.450913 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" event={"ID":"be580dcd-1367-419d-be4f-0b36c258fd13","Type":"ContainerDied","Data":"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.450940 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" event={"ID":"be580dcd-1367-419d-be4f-0b36c258fd13","Type":"ContainerDied","Data":"41dc924481a875296642d5181a6dc44913b21b365849b2296a37a67414248c80"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.450960 4727 scope.go:117] "RemoveContainer" containerID="759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.451102 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jndcv" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.455012 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cx7d2" event={"ID":"f16d06b5-d52a-48bb-8809-f06e0959ea6e","Type":"ContainerStarted","Data":"ceb803bf985e14d6f4b3ec771fd470023b0ddfe62cdf8230d39118d07ac4c194"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.456054 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be580dcd-1367-419d-be4f-0b36c258fd13" (UID: "be580dcd-1367-419d-be4f-0b36c258fd13"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.466622 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" event={"ID":"ede45115-5db7-4287-a93d-f886b0c7fbf5","Type":"ContainerStarted","Data":"c85608b2f1479f7085eb2d1eb78fff100ee3df3c83b24d736e319ec2cdb6f5ff"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.468378 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerStarted","Data":"414ca8c023e12a35c9e960d6a63fb48b1432262aeebc2e928a35cba503d742dd"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.470204 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdzcd" event={"ID":"2c46375d-fe70-42a2-a59c-182dbfcd83a8","Type":"ContainerStarted","Data":"a757cb1ed7e49af3b555f7e26b262fe1f4fb4bc0f7201a54077dd1afccef60a5"} Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.471288 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "be580dcd-1367-419d-be4f-0b36c258fd13" (UID: "be580dcd-1367-419d-be4f-0b36c258fd13"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.478174 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config" (OuterVolumeSpecName: "config") pod "be580dcd-1367-419d-be4f-0b36c258fd13" (UID: "be580dcd-1367-419d-be4f-0b36c258fd13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.489876 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "be580dcd-1367-419d-be4f-0b36c258fd13" (UID: "be580dcd-1367-419d-be4f-0b36c258fd13"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.490696 4727 scope.go:117] "RemoveContainer" containerID="b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.494634 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.494672 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.494683 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.494692 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be580dcd-1367-419d-be4f-0b36c258fd13-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.494705 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzwlr\" (UniqueName: \"kubernetes.io/projected/be580dcd-1367-419d-be4f-0b36c258fd13-kube-api-access-pzwlr\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.595580 4727 scope.go:117] "RemoveContainer" containerID="759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70" Sep 29 10:40:40 crc kubenswrapper[4727]: E0929 10:40:40.599626 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70\": container with ID starting with 759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70 not found: ID does not exist" containerID="759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.599689 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70"} err="failed to get container status \"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70\": rpc error: code = NotFound desc = could not find container \"759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70\": container with ID starting with 759dae922b45521145df5f54086ab549bf066c15e4c73b7ab366009dd1e7cf70 not found: ID does not exist" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.599728 4727 scope.go:117] "RemoveContainer" containerID="b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f" Sep 29 10:40:40 crc kubenswrapper[4727]: E0929 10:40:40.604612 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f\": container with ID starting with b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f not found: ID does not exist" containerID="b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.604694 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f"} err="failed to get container status \"b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f\": rpc error: code = NotFound desc = could not find container \"b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f\": container with ID starting with b603fb9f62ebe7d5d5ad36747b79fa1aa954a7a853db9400e456e8ff07550b2f not found: ID does not exist" Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.617484 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.764559 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:40 crc kubenswrapper[4727]: W0929 10:40:40.772871 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb25be5ca_0503_4333_b55b_1dc928fd04dd.slice/crio-71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d WatchSource:0}: Error finding container 71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d: Status 404 returned error can't find the container with id 71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.800752 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:40 crc kubenswrapper[4727]: I0929 10:40:40.810763 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jndcv"] Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.138749 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" path="/var/lib/kubelet/pods/be580dcd-1367-419d-be4f-0b36c258fd13/volumes" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.322285 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.352667 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7a6a9dee-ca84-4c2d-abc9-613bccd90764-etc-swift\") pod \"swift-storage-0\" (UID: \"7a6a9dee-ca84-4c2d-abc9-613bccd90764\") " pod="openstack/swift-storage-0" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.494767 4727 generic.go:334] "Generic (PLEG): container finished" podID="ede45115-5db7-4287-a93d-f886b0c7fbf5" containerID="33f40d2c6d65e7ee01b48b5b7ea6d6942b69f6a99ae4edfcaaf3477b21ab7333" exitCode=0 Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.494860 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" event={"ID":"ede45115-5db7-4287-a93d-f886b0c7fbf5","Type":"ContainerDied","Data":"33f40d2c6d65e7ee01b48b5b7ea6d6942b69f6a99ae4edfcaaf3477b21ab7333"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.503634 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdzcd" event={"ID":"2c46375d-fe70-42a2-a59c-182dbfcd83a8","Type":"ContainerStarted","Data":"7d2d0fdfae3d791d6f001d70e5b1e8c4547e642a70a19dbd15b5314a2b9d92d5"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.508293 4727 generic.go:334] "Generic (PLEG): container finished" podID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerID="f6d8695459b715a3428091ccc94dc8da5976ee2ee61dcbfdf77916c3a3e7a864" exitCode=0 Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.508385 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" event={"ID":"82a82880-cb7a-40e1-9908-03f142f6ff71","Type":"ContainerDied","Data":"f6d8695459b715a3428091ccc94dc8da5976ee2ee61dcbfdf77916c3a3e7a864"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.508414 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" event={"ID":"82a82880-cb7a-40e1-9908-03f142f6ff71","Type":"ContainerStarted","Data":"53a1d0c9fca2d98abe2fd625c68cde7901dbe57fda1f164ebda3ad4c4594d139"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.513662 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerStarted","Data":"71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.525314 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerStarted","Data":"234b53d9a1ebbdfb89be8c5a1f8c7a2ecf75c59b606c8210a9979cd11b9551e0"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.525425 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerStarted","Data":"5292fb802ca6f9af7b2318f2c8704b8753a5ea7c4cd56f16933096cbf32c04da"} Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.576143 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.583752 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mdzcd" podStartSLOduration=3.583733935 podStartE2EDuration="3.583733935s" podCreationTimestamp="2025-09-29 10:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:41.57617469 +0000 UTC m=+1111.749488052" watchObservedRunningTime="2025-09-29 10:40:41.583733935 +0000 UTC m=+1111.757047297" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.856654 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.932938 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc\") pod \"ede45115-5db7-4287-a93d-f886b0c7fbf5\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.933026 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb\") pod \"ede45115-5db7-4287-a93d-f886b0c7fbf5\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.933098 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbjh7\" (UniqueName: \"kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7\") pod \"ede45115-5db7-4287-a93d-f886b0c7fbf5\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.933119 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config\") pod \"ede45115-5db7-4287-a93d-f886b0c7fbf5\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.933237 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb\") pod \"ede45115-5db7-4287-a93d-f886b0c7fbf5\" (UID: \"ede45115-5db7-4287-a93d-f886b0c7fbf5\") " Sep 29 10:40:41 crc kubenswrapper[4727]: I0929 10:40:41.947075 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7" (OuterVolumeSpecName: "kube-api-access-zbjh7") pod "ede45115-5db7-4287-a93d-f886b0c7fbf5" (UID: "ede45115-5db7-4287-a93d-f886b0c7fbf5"). InnerVolumeSpecName "kube-api-access-zbjh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.040778 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbjh7\" (UniqueName: \"kubernetes.io/projected/ede45115-5db7-4287-a93d-f886b0c7fbf5-kube-api-access-zbjh7\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.065540 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ede45115-5db7-4287-a93d-f886b0c7fbf5" (UID: "ede45115-5db7-4287-a93d-f886b0c7fbf5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.087071 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ede45115-5db7-4287-a93d-f886b0c7fbf5" (UID: "ede45115-5db7-4287-a93d-f886b0c7fbf5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.093162 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ede45115-5db7-4287-a93d-f886b0c7fbf5" (UID: "ede45115-5db7-4287-a93d-f886b0c7fbf5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.117391 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config" (OuterVolumeSpecName: "config") pod "ede45115-5db7-4287-a93d-f886b0c7fbf5" (UID: "ede45115-5db7-4287-a93d-f886b0c7fbf5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.142820 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.142852 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.142862 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.142873 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ede45115-5db7-4287-a93d-f886b0c7fbf5-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.277139 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 29 10:40:42 crc kubenswrapper[4727]: W0929 10:40:42.279271 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a6a9dee_ca84_4c2d_abc9_613bccd90764.slice/crio-d6c18908c57d7f1750f0d1695a53545319daabaf23742759b3461d739b58153d WatchSource:0}: Error finding container d6c18908c57d7f1750f0d1695a53545319daabaf23742759b3461d739b58153d: Status 404 returned error can't find the container with id d6c18908c57d7f1750f0d1695a53545319daabaf23742759b3461d739b58153d Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.552237 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" event={"ID":"82a82880-cb7a-40e1-9908-03f142f6ff71","Type":"ContainerStarted","Data":"066b43bb1594b6e4a204ccd6714333c0c1e6cd6ba474b75b7da7e62b416f6c26"} Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.556298 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.562618 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"d6c18908c57d7f1750f0d1695a53545319daabaf23742759b3461d739b58153d"} Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.583946 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" podStartSLOduration=3.583915987 podStartE2EDuration="3.583915987s" podCreationTimestamp="2025-09-29 10:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:42.572942565 +0000 UTC m=+1112.746255957" watchObservedRunningTime="2025-09-29 10:40:42.583915987 +0000 UTC m=+1112.757229349" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.590610 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerStarted","Data":"6dcc88aeb582b1378f9505c78dea27c8a43e5e2b9d8985064f7dc746630758d1"} Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.608752 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerStarted","Data":"03b5cbc07f15306c9e65344f8a3914a5ab4e7b2e7bef9c874db6035b550228da"} Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.622421 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.622832 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-t4wzb" event={"ID":"ede45115-5db7-4287-a93d-f886b0c7fbf5","Type":"ContainerDied","Data":"c85608b2f1479f7085eb2d1eb78fff100ee3df3c83b24d736e319ec2cdb6f5ff"} Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.622886 4727 scope.go:117] "RemoveContainer" containerID="33f40d2c6d65e7ee01b48b5b7ea6d6942b69f6a99ae4edfcaaf3477b21ab7333" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.631237 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.6312148440000005 podStartE2EDuration="4.631214844s" podCreationTimestamp="2025-09-29 10:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:42.630917747 +0000 UTC m=+1112.804231119" watchObservedRunningTime="2025-09-29 10:40:42.631214844 +0000 UTC m=+1112.804528206" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.707303 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.712532 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-t4wzb"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779034 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2f68-account-create-9qmm6"] Sep 29 10:40:42 crc kubenswrapper[4727]: E0929 10:40:42.779518 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="init" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779532 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="init" Sep 29 10:40:42 crc kubenswrapper[4727]: E0929 10:40:42.779551 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="dnsmasq-dns" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779557 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="dnsmasq-dns" Sep 29 10:40:42 crc kubenswrapper[4727]: E0929 10:40:42.779571 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede45115-5db7-4287-a93d-f886b0c7fbf5" containerName="init" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779579 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede45115-5db7-4287-a93d-f886b0c7fbf5" containerName="init" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779729 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="be580dcd-1367-419d-be4f-0b36c258fd13" containerName="dnsmasq-dns" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.779747 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede45115-5db7-4287-a93d-f886b0c7fbf5" containerName="init" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.780387 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.788912 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.807666 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2f68-account-create-9qmm6"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.884718 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlwt\" (UniqueName: \"kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt\") pod \"barbican-2f68-account-create-9qmm6\" (UID: \"9ab205cd-2a27-40df-a1ca-2698379a1148\") " pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.910291 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.923138 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.973576 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7ab5-account-create-bfxmc"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.974716 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.977020 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.982294 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7ab5-account-create-bfxmc"] Sep 29 10:40:42 crc kubenswrapper[4727]: I0929 10:40:42.992824 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlwt\" (UniqueName: \"kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt\") pod \"barbican-2f68-account-create-9qmm6\" (UID: \"9ab205cd-2a27-40df-a1ca-2698379a1148\") " pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.009229 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.010769 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.056180 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.072998 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlwt\" (UniqueName: \"kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt\") pod \"barbican-2f68-account-create-9qmm6\" (UID: \"9ab205cd-2a27-40df-a1ca-2698379a1148\") " pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.094685 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.094965 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.095093 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtcz2\" (UniqueName: \"kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.095230 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpd9c\" (UniqueName: \"kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c\") pod \"cinder-7ab5-account-create-bfxmc\" (UID: \"9030b15b-1cf7-4797-ae8e-b6b0a6137b92\") " pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.095378 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.095485 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.117471 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.145106 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ede45115-5db7-4287-a93d-f886b0c7fbf5" path="/var/lib/kubelet/pods/ede45115-5db7-4287-a93d-f886b0c7fbf5/volumes" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.156494 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.185121 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.192472 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bdc4-account-create-7rr6z"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.193737 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.198632 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.198904 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.199001 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtcz2\" (UniqueName: \"kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.199100 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpd9c\" (UniqueName: \"kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c\") pod \"cinder-7ab5-account-create-bfxmc\" (UID: \"9030b15b-1cf7-4797-ae8e-b6b0a6137b92\") " pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.199175 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.199264 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.199371 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.200096 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.200662 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.201415 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.223027 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.223463 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bdc4-account-create-7rr6z"] Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.225225 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpd9c\" (UniqueName: \"kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c\") pod \"cinder-7ab5-account-create-bfxmc\" (UID: \"9030b15b-1cf7-4797-ae8e-b6b0a6137b92\") " pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.230161 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtcz2\" (UniqueName: \"kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2\") pod \"horizon-65fc4fcb97-6gnq7\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.301258 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfrft\" (UniqueName: \"kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft\") pod \"neutron-bdc4-account-create-7rr6z\" (UID: \"9b1ec441-5723-4541-8614-435ec4c53b07\") " pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.342792 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.395806 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.405270 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfrft\" (UniqueName: \"kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft\") pod \"neutron-bdc4-account-create-7rr6z\" (UID: \"9b1ec441-5723-4541-8614-435ec4c53b07\") " pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:43 crc kubenswrapper[4727]: I0929 10:40:43.429204 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfrft\" (UniqueName: \"kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft\") pod \"neutron-bdc4-account-create-7rr6z\" (UID: \"9b1ec441-5723-4541-8614-435ec4c53b07\") " pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.645193 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-log" containerID="cri-o://6dcc88aeb582b1378f9505c78dea27c8a43e5e2b9d8985064f7dc746630758d1" gracePeriod=30 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.645719 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerStarted","Data":"a4a67158a9018c3c839f8be630f61699778ccc2313ec80bda6c820a1124b40af"} Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.646028 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-httpd" containerID="cri-o://a4a67158a9018c3c839f8be630f61699778ccc2313ec80bda6c820a1124b40af" gracePeriod=30 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.649150 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.700986 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2f68-account-create-9qmm6"] Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.702559 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.702535578 podStartE2EDuration="4.702535578s" podCreationTimestamp="2025-09-29 10:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:43.673797559 +0000 UTC m=+1113.847110921" watchObservedRunningTime="2025-09-29 10:40:43.702535578 +0000 UTC m=+1113.875848940" Sep 29 10:40:44 crc kubenswrapper[4727]: W0929 10:40:43.726258 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab205cd_2a27_40df_a1ca_2698379a1148.slice/crio-b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d WatchSource:0}: Error finding container b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d: Status 404 returned error can't find the container with id b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:43.951646 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7ab5-account-create-bfxmc"] Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.674630 4727 generic.go:334] "Generic (PLEG): container finished" podID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerID="a4a67158a9018c3c839f8be630f61699778ccc2313ec80bda6c820a1124b40af" exitCode=0 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.674665 4727 generic.go:334] "Generic (PLEG): container finished" podID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerID="6dcc88aeb582b1378f9505c78dea27c8a43e5e2b9d8985064f7dc746630758d1" exitCode=143 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.674717 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerDied","Data":"a4a67158a9018c3c839f8be630f61699778ccc2313ec80bda6c820a1124b40af"} Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.674762 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerDied","Data":"6dcc88aeb582b1378f9505c78dea27c8a43e5e2b9d8985064f7dc746630758d1"} Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.686094 4727 generic.go:334] "Generic (PLEG): container finished" podID="9ab205cd-2a27-40df-a1ca-2698379a1148" containerID="41d2eb378ad10a0657fc3ccafb5692d2abafc83e5b33dda6ea43161e8a49c698" exitCode=0 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.686146 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2f68-account-create-9qmm6" event={"ID":"9ab205cd-2a27-40df-a1ca-2698379a1148","Type":"ContainerDied","Data":"41d2eb378ad10a0657fc3ccafb5692d2abafc83e5b33dda6ea43161e8a49c698"} Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.686195 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2f68-account-create-9qmm6" event={"ID":"9ab205cd-2a27-40df-a1ca-2698379a1148","Type":"ContainerStarted","Data":"b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d"} Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.686292 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-log" containerID="cri-o://234b53d9a1ebbdfb89be8c5a1f8c7a2ecf75c59b606c8210a9979cd11b9551e0" gracePeriod=30 Sep 29 10:40:44 crc kubenswrapper[4727]: I0929 10:40:44.686414 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-httpd" containerID="cri-o://03b5cbc07f15306c9e65344f8a3914a5ab4e7b2e7bef9c874db6035b550228da" gracePeriod=30 Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.280633 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bdc4-account-create-7rr6z"] Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.289795 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.694310 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ab5-account-create-bfxmc" event={"ID":"9030b15b-1cf7-4797-ae8e-b6b0a6137b92","Type":"ContainerStarted","Data":"d401c316ef483502f8ccd0808abbedd12aa46808a89ef9413e02448b36324477"} Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.702167 4727 generic.go:334] "Generic (PLEG): container finished" podID="23818aef-5aaf-4848-af35-64585311f0b1" containerID="03b5cbc07f15306c9e65344f8a3914a5ab4e7b2e7bef9c874db6035b550228da" exitCode=0 Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.702216 4727 generic.go:334] "Generic (PLEG): container finished" podID="23818aef-5aaf-4848-af35-64585311f0b1" containerID="234b53d9a1ebbdfb89be8c5a1f8c7a2ecf75c59b606c8210a9979cd11b9551e0" exitCode=143 Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.702227 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerDied","Data":"03b5cbc07f15306c9e65344f8a3914a5ab4e7b2e7bef9c874db6035b550228da"} Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.702287 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerDied","Data":"234b53d9a1ebbdfb89be8c5a1f8c7a2ecf75c59b606c8210a9979cd11b9551e0"} Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.704137 4727 generic.go:334] "Generic (PLEG): container finished" podID="2c46375d-fe70-42a2-a59c-182dbfcd83a8" containerID="7d2d0fdfae3d791d6f001d70e5b1e8c4547e642a70a19dbd15b5314a2b9d92d5" exitCode=0 Sep 29 10:40:45 crc kubenswrapper[4727]: I0929 10:40:45.704238 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdzcd" event={"ID":"2c46375d-fe70-42a2-a59c-182dbfcd83a8","Type":"ContainerDied","Data":"7d2d0fdfae3d791d6f001d70e5b1e8c4547e642a70a19dbd15b5314a2b9d92d5"} Sep 29 10:40:46 crc kubenswrapper[4727]: W0929 10:40:46.832809 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0b648b7_c07b_4516_9365_3b7680519463.slice/crio-8c53b925fe4c11ede16977b8599099dbf049e124e6f04c5daa60e792563a9afd WatchSource:0}: Error finding container 8c53b925fe4c11ede16977b8599099dbf049e124e6f04c5daa60e792563a9afd: Status 404 returned error can't find the container with id 8c53b925fe4c11ede16977b8599099dbf049e124e6f04c5daa60e792563a9afd Sep 29 10:40:47 crc kubenswrapper[4727]: I0929 10:40:47.722786 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerStarted","Data":"8c53b925fe4c11ede16977b8599099dbf049e124e6f04c5daa60e792563a9afd"} Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.246488 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.246828 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.479063 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.520188 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.521723 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.526394 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.541564 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.606996 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.636895 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54cbdfc6b8-xl7cd"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.638429 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639249 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639451 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639497 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639528 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639555 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639663 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktrz5\" (UniqueName: \"kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.639693 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.647505 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54cbdfc6b8-xl7cd"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.734232 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741102 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741199 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741234 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-scripts\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741280 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-combined-ca-bundle\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741353 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741385 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741417 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741443 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741528 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhz9b\" (UniqueName: \"kubernetes.io/projected/4b7b8921-ae97-45b0-a082-07db49561514-kube-api-access-jhz9b\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741580 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-secret-key\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741602 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-tls-certs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741645 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7b8921-ae97-45b0-a082-07db49561514-logs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741681 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-config-data\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.741713 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktrz5\" (UniqueName: \"kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.742612 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.743290 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.743512 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.749923 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.750090 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.751650 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.789006 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktrz5\" (UniqueName: \"kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5\") pod \"horizon-559c5978cd-jb7bp\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.793487 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.793712 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-sknk6" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="dnsmasq-dns" containerID="cri-o://b3834598a41238f8a7826bd8e8245b873f2d6bbaad9462770c5c1d0ccc4e3928" gracePeriod=10 Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846572 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-scripts\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846650 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-combined-ca-bundle\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846782 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhz9b\" (UniqueName: \"kubernetes.io/projected/4b7b8921-ae97-45b0-a082-07db49561514-kube-api-access-jhz9b\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846834 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-secret-key\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846855 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-tls-certs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846894 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7b8921-ae97-45b0-a082-07db49561514-logs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.846927 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-config-data\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.848827 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-config-data\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.849321 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b7b8921-ae97-45b0-a082-07db49561514-scripts\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.854193 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7b8921-ae97-45b0-a082-07db49561514-logs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.854477 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-secret-key\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.857811 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.858546 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-horizon-tls-certs\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.870960 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7b8921-ae97-45b0-a082-07db49561514-combined-ca-bundle\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.873191 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhz9b\" (UniqueName: \"kubernetes.io/projected/4b7b8921-ae97-45b0-a082-07db49561514-kube-api-access-jhz9b\") pod \"horizon-54cbdfc6b8-xl7cd\" (UID: \"4b7b8921-ae97-45b0-a082-07db49561514\") " pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:49 crc kubenswrapper[4727]: I0929 10:40:49.956255 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:50 crc kubenswrapper[4727]: I0929 10:40:50.757362 4727 generic.go:334] "Generic (PLEG): container finished" podID="4460bc15-e725-4559-ab19-344e52088e27" containerID="b3834598a41238f8a7826bd8e8245b873f2d6bbaad9462770c5c1d0ccc4e3928" exitCode=0 Sep 29 10:40:50 crc kubenswrapper[4727]: I0929 10:40:50.757457 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sknk6" event={"ID":"4460bc15-e725-4559-ab19-344e52088e27","Type":"ContainerDied","Data":"b3834598a41238f8a7826bd8e8245b873f2d6bbaad9462770c5c1d0ccc4e3928"} Sep 29 10:40:51 crc kubenswrapper[4727]: I0929 10:40:51.764144 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-sknk6" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.665618 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.666366 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n695h58fh95h655h76h66fh649h56bh594h595h545h5ffh54bh557hfch55fh67bh654h5c4hc9h5ffh5bch665h655h598h676h85h5b5h7fh5d6h56bh66q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8xnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-56b7b5f9ff-9bp57_openstack(88e12d3e-6d5c-4c08-9279-cdcd18108f1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.669258 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-56b7b5f9ff-9bp57" podUID="88e12d3e-6d5c-4c08-9279-cdcd18108f1e" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.697333 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.697560 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n664h565h54dh684h8h684h698h545h55bh68dh566hf8hfh564h5b7h677h74hc4h5b9hc7h677h9chdbh5f9hb4h585h56bh54dhcch54h5b4h68dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rksws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-64d84c44f5-8xbxm_openstack(7df9aec9-2126-49b0-a4ec-7dd62cf01ac4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:40:54 crc kubenswrapper[4727]: E0929 10:40:54.741765 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-64d84c44f5-8xbxm" podUID="7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.791850 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b25be5ca-0503-4333-b55b-1dc928fd04dd","Type":"ContainerDied","Data":"71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d"} Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.792062 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71f36e9c010a0e4e44566745cbf442d459c6ed111ab5e6d71468aa4b4b5d516d" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.793779 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2f68-account-create-9qmm6" event={"ID":"9ab205cd-2a27-40df-a1ca-2698379a1148","Type":"ContainerDied","Data":"b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d"} Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.793805 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b554891858d14762a4b5a669795eda438807e2be0dcf39bd8f25653197b75a7d" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.795460 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdzcd" event={"ID":"2c46375d-fe70-42a2-a59c-182dbfcd83a8","Type":"ContainerDied","Data":"a757cb1ed7e49af3b555f7e26b262fe1f4fb4bc0f7201a54077dd1afccef60a5"} Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.795519 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a757cb1ed7e49af3b555f7e26b262fe1f4fb4bc0f7201a54077dd1afccef60a5" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.796833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bdc4-account-create-7rr6z" event={"ID":"9b1ec441-5723-4541-8614-435ec4c53b07","Type":"ContainerStarted","Data":"75697c266b7ddbbf7785a326ff6a597a81df8646d05de7789660b11aa656d42b"} Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.838856 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.842046 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.843685 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.931931 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932040 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932072 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932112 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjcb5\" (UniqueName: \"kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932136 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932186 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932217 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932551 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932582 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlwt\" (UniqueName: \"kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt\") pod \"9ab205cd-2a27-40df-a1ca-2698379a1148\" (UID: \"9ab205cd-2a27-40df-a1ca-2698379a1148\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932770 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932842 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932908 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.932947 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"b25be5ca-0503-4333-b55b-1dc928fd04dd\" (UID: \"b25be5ca-0503-4333-b55b-1dc928fd04dd\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.933576 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8r8h\" (UniqueName: \"kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.933669 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys\") pod \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\" (UID: \"2c46375d-fe70-42a2-a59c-182dbfcd83a8\") " Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.934291 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs" (OuterVolumeSpecName: "logs") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.934644 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.934667 4727 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b25be5ca-0503-4333-b55b-1dc928fd04dd-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.939282 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt" (OuterVolumeSpecName: "kube-api-access-tqlwt") pod "9ab205cd-2a27-40df-a1ca-2698379a1148" (UID: "9ab205cd-2a27-40df-a1ca-2698379a1148"). InnerVolumeSpecName "kube-api-access-tqlwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.939949 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.944047 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts" (OuterVolumeSpecName: "scripts") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.944164 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.944462 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h" (OuterVolumeSpecName: "kube-api-access-b8r8h") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "kube-api-access-b8r8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.953205 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.957299 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5" (OuterVolumeSpecName: "kube-api-access-jjcb5") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "kube-api-access-jjcb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.974266 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts" (OuterVolumeSpecName: "scripts") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.979417 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:54 crc kubenswrapper[4727]: I0929 10:40:54.990790 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.002651 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data" (OuterVolumeSpecName: "config-data") pod "2c46375d-fe70-42a2-a59c-182dbfcd83a8" (UID: "2c46375d-fe70-42a2-a59c-182dbfcd83a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.003583 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data" (OuterVolumeSpecName: "config-data") pod "b25be5ca-0503-4333-b55b-1dc928fd04dd" (UID: "b25be5ca-0503-4333-b55b-1dc928fd04dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035845 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035878 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjcb5\" (UniqueName: \"kubernetes.io/projected/b25be5ca-0503-4333-b55b-1dc928fd04dd-kube-api-access-jjcb5\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035889 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035900 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035908 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035916 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlwt\" (UniqueName: \"kubernetes.io/projected/9ab205cd-2a27-40df-a1ca-2698379a1148-kube-api-access-tqlwt\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035924 4727 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035931 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035965 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035974 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8r8h\" (UniqueName: \"kubernetes.io/projected/2c46375d-fe70-42a2-a59c-182dbfcd83a8-kube-api-access-b8r8h\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035983 4727 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2c46375d-fe70-42a2-a59c-182dbfcd83a8-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.035991 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25be5ca-0503-4333-b55b-1dc928fd04dd-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.055281 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.137696 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: E0929 10:40:55.294017 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Sep 29 10:40:55 crc kubenswrapper[4727]: E0929 10:40:55.294194 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n547hcch5ffh77h94h79hcbh5d8hc4h54h556h688h684h66dh9bh674hch5h54fh5f5h8ch599h9fh5d4h7ch695h57h5c7h688h9hd4h5c9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nphtd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(12d4eb15-f153-4824-85a4-40bceb032c7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.294465 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.303116 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.312197 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.318762 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341024 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rksws\" (UniqueName: \"kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws\") pod \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341074 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key\") pod \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341126 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc\") pod \"4460bc15-e725-4559-ab19-344e52088e27\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341149 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crl2q\" (UniqueName: \"kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q\") pod \"4460bc15-e725-4559-ab19-344e52088e27\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341176 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config\") pod \"4460bc15-e725-4559-ab19-344e52088e27\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341203 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341220 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69bdl\" (UniqueName: \"kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341252 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts\") pod \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341277 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts\") pod \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341298 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341313 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data\") pod \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341362 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key\") pod \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341404 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8xnt\" (UniqueName: \"kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt\") pod \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341462 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.341490 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.342251 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data\") pod \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.342445 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.342587 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs\") pod \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\" (UID: \"88e12d3e-6d5c-4c08-9279-cdcd18108f1e\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.342739 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs\") pod \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\" (UID: \"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.342769 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb\") pod \"4460bc15-e725-4559-ab19-344e52088e27\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.343111 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs" (OuterVolumeSpecName: "logs") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.343090 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb\") pod \"4460bc15-e725-4559-ab19-344e52088e27\" (UID: \"4460bc15-e725-4559-ab19-344e52088e27\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.344147 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"23818aef-5aaf-4848-af35-64585311f0b1\" (UID: \"23818aef-5aaf-4848-af35-64585311f0b1\") " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.345294 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.346765 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data" (OuterVolumeSpecName: "config-data") pod "88e12d3e-6d5c-4c08-9279-cdcd18108f1e" (UID: "88e12d3e-6d5c-4c08-9279-cdcd18108f1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.349414 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts" (OuterVolumeSpecName: "scripts") pod "88e12d3e-6d5c-4c08-9279-cdcd18108f1e" (UID: "88e12d3e-6d5c-4c08-9279-cdcd18108f1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.349793 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.350068 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts" (OuterVolumeSpecName: "scripts") pod "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" (UID: "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.351367 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs" (OuterVolumeSpecName: "logs") pod "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" (UID: "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.351617 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs" (OuterVolumeSpecName: "logs") pod "88e12d3e-6d5c-4c08-9279-cdcd18108f1e" (UID: "88e12d3e-6d5c-4c08-9279-cdcd18108f1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.353754 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data" (OuterVolumeSpecName: "config-data") pod "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" (UID: "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.353933 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.354352 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt" (OuterVolumeSpecName: "kube-api-access-g8xnt") pod "88e12d3e-6d5c-4c08-9279-cdcd18108f1e" (UID: "88e12d3e-6d5c-4c08-9279-cdcd18108f1e"). InnerVolumeSpecName "kube-api-access-g8xnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.368310 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts" (OuterVolumeSpecName: "scripts") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.368323 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q" (OuterVolumeSpecName: "kube-api-access-crl2q") pod "4460bc15-e725-4559-ab19-344e52088e27" (UID: "4460bc15-e725-4559-ab19-344e52088e27"). InnerVolumeSpecName "kube-api-access-crl2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.368409 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "88e12d3e-6d5c-4c08-9279-cdcd18108f1e" (UID: "88e12d3e-6d5c-4c08-9279-cdcd18108f1e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.372904 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl" (OuterVolumeSpecName: "kube-api-access-69bdl") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "kube-api-access-69bdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.375885 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws" (OuterVolumeSpecName: "kube-api-access-rksws") pod "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" (UID: "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4"). InnerVolumeSpecName "kube-api-access-rksws". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.375902 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" (UID: "7df9aec9-2126-49b0-a4ec-7dd62cf01ac4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.421284 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.430670 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4460bc15-e725-4559-ab19-344e52088e27" (UID: "4460bc15-e725-4559-ab19-344e52088e27"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447264 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447291 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crl2q\" (UniqueName: \"kubernetes.io/projected/4460bc15-e725-4559-ab19-344e52088e27-kube-api-access-crl2q\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447301 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447311 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69bdl\" (UniqueName: \"kubernetes.io/projected/23818aef-5aaf-4848-af35-64585311f0b1-kube-api-access-69bdl\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447325 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447347 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447357 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447368 4727 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447377 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8xnt\" (UniqueName: \"kubernetes.io/projected/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-kube-api-access-g8xnt\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447385 4727 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23818aef-5aaf-4848-af35-64585311f0b1-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447392 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447401 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447410 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447420 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447441 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447451 4727 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88e12d3e-6d5c-4c08-9279-cdcd18108f1e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447460 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rksws\" (UniqueName: \"kubernetes.io/projected/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4-kube-api-access-rksws\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447564 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config" (OuterVolumeSpecName: "config") pod "4460bc15-e725-4559-ab19-344e52088e27" (UID: "4460bc15-e725-4559-ab19-344e52088e27"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.447752 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4460bc15-e725-4559-ab19-344e52088e27" (UID: "4460bc15-e725-4559-ab19-344e52088e27"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.450684 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4460bc15-e725-4559-ab19-344e52088e27" (UID: "4460bc15-e725-4559-ab19-344e52088e27"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.466613 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data" (OuterVolumeSpecName: "config-data") pod "23818aef-5aaf-4848-af35-64585311f0b1" (UID: "23818aef-5aaf-4848-af35-64585311f0b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.485148 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.550206 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23818aef-5aaf-4848-af35-64585311f0b1-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.550593 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.550608 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.550621 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.550633 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460bc15-e725-4559-ab19-344e52088e27-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.807260 4727 generic.go:334] "Generic (PLEG): container finished" podID="9b1ec441-5723-4541-8614-435ec4c53b07" containerID="6512a5663c0429b948b72b58ed9346f051b8ec842852e16c408b2f4d9b965451" exitCode=0 Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.807534 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bdc4-account-create-7rr6z" event={"ID":"9b1ec441-5723-4541-8614-435ec4c53b07","Type":"ContainerDied","Data":"6512a5663c0429b948b72b58ed9346f051b8ec842852e16c408b2f4d9b965451"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.815827 4727 generic.go:334] "Generic (PLEG): container finished" podID="9030b15b-1cf7-4797-ae8e-b6b0a6137b92" containerID="05dd1927ee7df3a348122fc86cee1eac5074b30ee915edf9ddc7d0943d823b89" exitCode=0 Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.815911 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ab5-account-create-bfxmc" event={"ID":"9030b15b-1cf7-4797-ae8e-b6b0a6137b92","Type":"ContainerDied","Data":"05dd1927ee7df3a348122fc86cee1eac5074b30ee915edf9ddc7d0943d823b89"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.817738 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d84c44f5-8xbxm" event={"ID":"7df9aec9-2126-49b0-a4ec-7dd62cf01ac4","Type":"ContainerDied","Data":"25fecaa7557d3c0dd2d3ff6f86c1820f71f0b8f30258a647fcfd9a95342f838f"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.817786 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d84c44f5-8xbxm" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.825231 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56b7b5f9ff-9bp57" event={"ID":"88e12d3e-6d5c-4c08-9279-cdcd18108f1e","Type":"ContainerDied","Data":"87e5fe3e1478f45e19aee2b7b5e0eb85931d5a51daf6530899cee672ecb972bf"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.825535 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56b7b5f9ff-9bp57" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.838861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"47df5f35ce3d96e7e02416899df3d8c394d974743e6038af538fc9249fefa3b6"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.847883 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cx7d2" event={"ID":"f16d06b5-d52a-48bb-8809-f06e0959ea6e","Type":"ContainerStarted","Data":"2cf4f9b396b3e7d89238c61a8c91e2589898f7c6cc75e004dd9467a094110a6f"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.860372 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.860809 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"23818aef-5aaf-4848-af35-64585311f0b1","Type":"ContainerDied","Data":"5292fb802ca6f9af7b2318f2c8704b8753a5ea7c4cd56f16933096cbf32c04da"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.861096 4727 scope.go:117] "RemoveContainer" containerID="03b5cbc07f15306c9e65344f8a3914a5ab4e7b2e7bef9c874db6035b550228da" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.863150 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerStarted","Data":"fa3b74a407f12888522d7e2dca80344ae4eb404a0123f5b27ed343722dda4ddd"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.880478 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sknk6" event={"ID":"4460bc15-e725-4559-ab19-344e52088e27","Type":"ContainerDied","Data":"0a94ecc3d55ac0e9bce676200c8533aa624afb67a077454d163e2de95ea554c9"} Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.880489 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdzcd" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.880562 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.880625 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sknk6" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.880526 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2f68-account-create-9qmm6" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.882461 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cx7d2" podStartSLOduration=2.789704255 podStartE2EDuration="17.88244529s" podCreationTimestamp="2025-09-29 10:40:38 +0000 UTC" firstStartedPulling="2025-09-29 10:40:40.306668087 +0000 UTC m=+1110.479981449" lastFinishedPulling="2025-09-29 10:40:55.399409122 +0000 UTC m=+1125.572722484" observedRunningTime="2025-09-29 10:40:55.868355917 +0000 UTC m=+1126.041669279" watchObservedRunningTime="2025-09-29 10:40:55.88244529 +0000 UTC m=+1126.055758652" Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.928499 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54cbdfc6b8-xl7cd"] Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.960352 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mdzcd"] Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.967724 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mdzcd"] Sep 29 10:40:55 crc kubenswrapper[4727]: W0929 10:40:55.972497 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b7b8921_ae97_45b0_a082_07db49561514.slice/crio-68d995b83c194ca22586cf1f3f13a2d4d7e7ff49123f0378bc2f23ce2be268b7 WatchSource:0}: Error finding container 68d995b83c194ca22586cf1f3f13a2d4d7e7ff49123f0378bc2f23ce2be268b7: Status 404 returned error can't find the container with id 68d995b83c194ca22586cf1f3f13a2d4d7e7ff49123f0378bc2f23ce2be268b7 Sep 29 10:40:55 crc kubenswrapper[4727]: I0929 10:40:55.993572 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.017388 4727 scope.go:117] "RemoveContainer" containerID="234b53d9a1ebbdfb89be8c5a1f8c7a2ecf75c59b606c8210a9979cd11b9551e0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.061845 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nh4r9"] Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062291 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab205cd-2a27-40df-a1ca-2698379a1148" containerName="mariadb-account-create" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062311 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab205cd-2a27-40df-a1ca-2698379a1148" containerName="mariadb-account-create" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062351 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062359 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062372 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062380 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062395 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062404 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062418 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c46375d-fe70-42a2-a59c-182dbfcd83a8" containerName="keystone-bootstrap" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062426 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c46375d-fe70-42a2-a59c-182dbfcd83a8" containerName="keystone-bootstrap" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062440 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062447 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062466 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="init" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062474 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="init" Sep 29 10:40:56 crc kubenswrapper[4727]: E0929 10:40:56.062483 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="dnsmasq-dns" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.062490 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="dnsmasq-dns" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065631 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065695 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c46375d-fe70-42a2-a59c-182dbfcd83a8" containerName="keystone-bootstrap" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065711 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065727 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4460bc15-e725-4559-ab19-344e52088e27" containerName="dnsmasq-dns" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065737 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" containerName="glance-log" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065749 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="23818aef-5aaf-4848-af35-64585311f0b1" containerName="glance-httpd" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.065766 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab205cd-2a27-40df-a1ca-2698379a1148" containerName="mariadb-account-create" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.066474 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.068873 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.069211 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.069585 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2g6jb" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.071763 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.074121 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nh4r9"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.106297 4727 scope.go:117] "RemoveContainer" containerID="b3834598a41238f8a7826bd8e8245b873f2d6bbaad9462770c5c1d0ccc4e3928" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165432 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165479 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165543 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165585 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cswvr\" (UniqueName: \"kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165609 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.165713 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.176496 4727 scope.go:117] "RemoveContainer" containerID="698dbe80bb166c232574ca754314c3dee83cd6a2acadda942b88789556e54f95" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.268630 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.268994 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.269018 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.269592 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.269758 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cswvr\" (UniqueName: \"kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.269800 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.279126 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.279231 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.290433 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.300185 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cswvr\" (UniqueName: \"kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.322935 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-56b7b5f9ff-9bp57"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.330795 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.332713 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.334833 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys\") pod \"keystone-bootstrap-nh4r9\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.355942 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.384169 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sknk6"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.426981 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.437375 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.447132 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.448779 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.453037 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.454671 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.454890 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vlvqq" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.459564 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.473979 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474302 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474406 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474469 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474539 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474592 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gqw7\" (UniqueName: \"kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474632 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.474668 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.521844 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.533692 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.549769 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575636 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gqw7\" (UniqueName: \"kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575716 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575745 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575764 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575844 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575957 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.575996 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.576050 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.577074 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.577261 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.578102 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.584122 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64d84c44f5-8xbxm"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.596550 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.600228 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.601565 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.604808 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gqw7\" (UniqueName: \"kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.613999 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.614770 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.641542 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.646916 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.653697 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.655732 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.657437 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.659095 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.659294 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.782468 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.785938 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786012 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786175 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786212 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786450 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786539 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786923 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.786957 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2npp8\" (UniqueName: \"kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889626 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889667 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889770 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889858 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889908 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.889971 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.890046 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2npp8\" (UniqueName: \"kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.891476 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.892157 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.892796 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.900458 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.901203 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.910342 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.912465 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.917377 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2npp8\" (UniqueName: \"kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.918753 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerStarted","Data":"6eb109f03a51bf37afc95e6079b2f74e3158c6e40ed77d59f590d891d0ea76c0"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.918916 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65fc4fcb97-6gnq7" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon-log" containerID="cri-o://fa3b74a407f12888522d7e2dca80344ae4eb404a0123f5b27ed343722dda4ddd" gracePeriod=30 Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.920290 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65fc4fcb97-6gnq7" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon" containerID="cri-o://6eb109f03a51bf37afc95e6079b2f74e3158c6e40ed77d59f590d891d0ea76c0" gracePeriod=30 Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.937955 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerStarted","Data":"4cd6244f38b2627ab0c262de29eecfcff1b39f6d175990a0f329ccd2991f7f30"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.938011 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerStarted","Data":"481952bf105362fb6dce800e1009a43adcb4406f1e2796fa433d14461722dab4"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.947606 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65fc4fcb97-6gnq7" podStartSLOduration=6.363132587 podStartE2EDuration="14.947574093s" podCreationTimestamp="2025-09-29 10:40:42 +0000 UTC" firstStartedPulling="2025-09-29 10:40:46.839149848 +0000 UTC m=+1117.012463210" lastFinishedPulling="2025-09-29 10:40:55.423591354 +0000 UTC m=+1125.596904716" observedRunningTime="2025-09-29 10:40:56.938046688 +0000 UTC m=+1127.111360050" watchObservedRunningTime="2025-09-29 10:40:56.947574093 +0000 UTC m=+1127.120887465" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.955628 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " pod="openstack/glance-default-external-api-0" Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.959029 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cbdfc6b8-xl7cd" event={"ID":"4b7b8921-ae97-45b0-a082-07db49561514","Type":"ContainerStarted","Data":"d30155005043b1b32aaaef5fbf887b92398b40f4976a4036a584a6684b892975"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.959079 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cbdfc6b8-xl7cd" event={"ID":"4b7b8921-ae97-45b0-a082-07db49561514","Type":"ContainerStarted","Data":"68d995b83c194ca22586cf1f3f13a2d4d7e7ff49123f0378bc2f23ce2be268b7"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.976134 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"c7f725d5e77675ad84069e05db6428384c40228774d05a162a357ac2b308374b"} Sep 29 10:40:56 crc kubenswrapper[4727]: I0929 10:40:56.976188 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"4902cd5b76ec280be92ffc414f40c3f6c571f72370d0a6ecb075981ff9c04f69"} Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.070483 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nh4r9"] Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.090487 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.120777 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23818aef-5aaf-4848-af35-64585311f0b1" path="/var/lib/kubelet/pods/23818aef-5aaf-4848-af35-64585311f0b1/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.121844 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c46375d-fe70-42a2-a59c-182dbfcd83a8" path="/var/lib/kubelet/pods/2c46375d-fe70-42a2-a59c-182dbfcd83a8/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.122429 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4460bc15-e725-4559-ab19-344e52088e27" path="/var/lib/kubelet/pods/4460bc15-e725-4559-ab19-344e52088e27/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.123670 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df9aec9-2126-49b0-a4ec-7dd62cf01ac4" path="/var/lib/kubelet/pods/7df9aec9-2126-49b0-a4ec-7dd62cf01ac4/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.124928 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88e12d3e-6d5c-4c08-9279-cdcd18108f1e" path="/var/lib/kubelet/pods/88e12d3e-6d5c-4c08-9279-cdcd18108f1e/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.125467 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b25be5ca-0503-4333-b55b-1dc928fd04dd" path="/var/lib/kubelet/pods/b25be5ca-0503-4333-b55b-1dc928fd04dd/volumes" Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.437871 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:40:57 crc kubenswrapper[4727]: I0929 10:40:57.704403 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:57.999958 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh4r9" event={"ID":"9fc77a68-5c6e-4325-9846-42b1e4f286ea","Type":"ContainerStarted","Data":"f08a74b68b3512eab0caf35f07e1f1e08a7a11fe07dc70518c49d4eb79cd8f20"} Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.000008 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh4r9" event={"ID":"9fc77a68-5c6e-4325-9846-42b1e4f286ea","Type":"ContainerStarted","Data":"8eb08975bcb6303a931ad027c1a315a5d142440712d4db1645bbf5450d793819"} Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.001763 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerStarted","Data":"d4e2464b9ae71a314ef444db324581c200bec588ddbb7b07b1c93c3d46fa4633"} Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.003597 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cbdfc6b8-xl7cd" event={"ID":"4b7b8921-ae97-45b0-a082-07db49561514","Type":"ContainerStarted","Data":"5fc0ab422e3b5947b5347822b9293d1829fcbb559c30662d02684d46bc996272"} Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.006068 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"7a45d84ede0ab70b71d3c8b0693832d65a041a52aad597586fd56416467bf621"} Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.023790 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-68lgd"] Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.025149 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.026450 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nh4r9" podStartSLOduration=2.026433921 podStartE2EDuration="2.026433921s" podCreationTimestamp="2025-09-29 10:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:58.024230554 +0000 UTC m=+1128.197543916" watchObservedRunningTime="2025-09-29 10:40:58.026433921 +0000 UTC m=+1128.199747283" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.026768 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-wx6p9" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.033149 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.044931 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-68lgd"] Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.064201 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-559c5978cd-jb7bp" podStartSLOduration=9.064149761 podStartE2EDuration="9.064149761s" podCreationTimestamp="2025-09-29 10:40:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:58.055499379 +0000 UTC m=+1128.228812741" watchObservedRunningTime="2025-09-29 10:40:58.064149761 +0000 UTC m=+1128.237463123" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.085814 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-54cbdfc6b8-xl7cd" podStartSLOduration=9.085797089 podStartE2EDuration="9.085797089s" podCreationTimestamp="2025-09-29 10:40:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:40:58.082117264 +0000 UTC m=+1128.255430626" watchObservedRunningTime="2025-09-29 10:40:58.085797089 +0000 UTC m=+1128.259110451" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.118934 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.119600 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc9pq\" (UniqueName: \"kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.119746 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: W0929 10:40:58.151937 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf00fce5b_a355_42e2_94c7_63810205536f.slice/crio-89acd8e1f6013b6d434f36580975c6c8da4d309201dc1dba406964b53d22ca93 WatchSource:0}: Error finding container 89acd8e1f6013b6d434f36580975c6c8da4d309201dc1dba406964b53d22ca93: Status 404 returned error can't find the container with id 89acd8e1f6013b6d434f36580975c6c8da4d309201dc1dba406964b53d22ca93 Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.222064 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc9pq\" (UniqueName: \"kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.222138 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.222212 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.237457 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.237829 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.242039 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc9pq\" (UniqueName: \"kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq\") pod \"barbican-db-sync-68lgd\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.313979 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.323617 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.347982 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-68lgd" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.424319 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpd9c\" (UniqueName: \"kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c\") pod \"9030b15b-1cf7-4797-ae8e-b6b0a6137b92\" (UID: \"9030b15b-1cf7-4797-ae8e-b6b0a6137b92\") " Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.424727 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfrft\" (UniqueName: \"kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft\") pod \"9b1ec441-5723-4541-8614-435ec4c53b07\" (UID: \"9b1ec441-5723-4541-8614-435ec4c53b07\") " Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.432820 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft" (OuterVolumeSpecName: "kube-api-access-bfrft") pod "9b1ec441-5723-4541-8614-435ec4c53b07" (UID: "9b1ec441-5723-4541-8614-435ec4c53b07"). InnerVolumeSpecName "kube-api-access-bfrft". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.432959 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c" (OuterVolumeSpecName: "kube-api-access-fpd9c") pod "9030b15b-1cf7-4797-ae8e-b6b0a6137b92" (UID: "9030b15b-1cf7-4797-ae8e-b6b0a6137b92"). InnerVolumeSpecName "kube-api-access-fpd9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.526506 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfrft\" (UniqueName: \"kubernetes.io/projected/9b1ec441-5723-4541-8614-435ec4c53b07-kube-api-access-bfrft\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.526937 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpd9c\" (UniqueName: \"kubernetes.io/projected/9030b15b-1cf7-4797-ae8e-b6b0a6137b92-kube-api-access-fpd9c\") on node \"crc\" DevicePath \"\"" Sep 29 10:40:58 crc kubenswrapper[4727]: I0929 10:40:58.872363 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-68lgd"] Sep 29 10:40:58 crc kubenswrapper[4727]: W0929 10:40:58.892486 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74fb3935_9e35_4a18_a2e4_b4442581bc00.slice/crio-b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554 WatchSource:0}: Error finding container b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554: Status 404 returned error can't find the container with id b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554 Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.034674 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-68lgd" event={"ID":"74fb3935-9e35-4a18-a2e4-b4442581bc00","Type":"ContainerStarted","Data":"b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.038290 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerStarted","Data":"4159dc13afebbe60f4ba315b81a3748e8cc5deb4542e1bec8aa38974b395c8a8"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.040845 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerStarted","Data":"8d8aa7de856cc6bf1639b3a983d2c3f34670e982a11e9d4d806252c2094a5c17"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.040901 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerStarted","Data":"89acd8e1f6013b6d434f36580975c6c8da4d309201dc1dba406964b53d22ca93"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.042774 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bdc4-account-create-7rr6z" event={"ID":"9b1ec441-5723-4541-8614-435ec4c53b07","Type":"ContainerDied","Data":"75697c266b7ddbbf7785a326ff6a597a81df8646d05de7789660b11aa656d42b"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.042807 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75697c266b7ddbbf7785a326ff6a597a81df8646d05de7789660b11aa656d42b" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.042868 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bdc4-account-create-7rr6z" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.049784 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ab5-account-create-bfxmc" event={"ID":"9030b15b-1cf7-4797-ae8e-b6b0a6137b92","Type":"ContainerDied","Data":"d401c316ef483502f8ccd0808abbedd12aa46808a89ef9413e02448b36324477"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.049833 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d401c316ef483502f8ccd0808abbedd12aa46808a89ef9413e02448b36324477" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.049832 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ab5-account-create-bfxmc" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.057124 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerStarted","Data":"3809597929b99fccd1f06ad19ec56e9c734668bea3069883bac20cdb27913ae7"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.057171 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerStarted","Data":"a87998dcd3dd24a025ca346bbd4cbdfc812e118d5e2c887dba3a2af244d265f2"} Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.858845 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.859419 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.956976 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:40:59 crc kubenswrapper[4727]: I0929 10:40:59.957062 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:41:00 crc kubenswrapper[4727]: I0929 10:41:00.084114 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerStarted","Data":"ae0b44851e2f70479e7adc332a042a7f624d3c636a668dba6b88fa26d8703144"} Sep 29 10:41:00 crc kubenswrapper[4727]: I0929 10:41:00.097823 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerStarted","Data":"a0a6dfca2b0543c83afb7c1f48c5eb590f228bdebf1a3522a00edaa1e3a0a59f"} Sep 29 10:41:00 crc kubenswrapper[4727]: I0929 10:41:00.122388 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.122368037 podStartE2EDuration="4.122368037s" podCreationTimestamp="2025-09-29 10:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:00.120469308 +0000 UTC m=+1130.293782680" watchObservedRunningTime="2025-09-29 10:41:00.122368037 +0000 UTC m=+1130.295681419" Sep 29 10:41:00 crc kubenswrapper[4727]: I0929 10:41:00.147937 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.147913654 podStartE2EDuration="4.147913654s" podCreationTimestamp="2025-09-29 10:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:00.143913492 +0000 UTC m=+1130.317226854" watchObservedRunningTime="2025-09-29 10:41:00.147913654 +0000 UTC m=+1130.321227017" Sep 29 10:41:01 crc kubenswrapper[4727]: I0929 10:41:01.159566 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"839557dbd0204da1d6cf11e3e53d624a49d9f1fa9d33da48c3cfa23a84b15aef"} Sep 29 10:41:01 crc kubenswrapper[4727]: I0929 10:41:01.159956 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"b467cfde464256feedd52b50e5c50eeb505dbaa1a780843548d99d90d8154a55"} Sep 29 10:41:01 crc kubenswrapper[4727]: I0929 10:41:01.159971 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"01cf84d6b5217c878a98686b31dc65e7d39083d5ba54cfef3ea808d2f33ffb4a"} Sep 29 10:41:01 crc kubenswrapper[4727]: I0929 10:41:01.163194 4727 generic.go:334] "Generic (PLEG): container finished" podID="f16d06b5-d52a-48bb-8809-f06e0959ea6e" containerID="2cf4f9b396b3e7d89238c61a8c91e2589898f7c6cc75e004dd9467a094110a6f" exitCode=0 Sep 29 10:41:01 crc kubenswrapper[4727]: I0929 10:41:01.164139 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cx7d2" event={"ID":"f16d06b5-d52a-48bb-8809-f06e0959ea6e","Type":"ContainerDied","Data":"2cf4f9b396b3e7d89238c61a8c91e2589898f7c6cc75e004dd9467a094110a6f"} Sep 29 10:41:02 crc kubenswrapper[4727]: I0929 10:41:02.174345 4727 generic.go:334] "Generic (PLEG): container finished" podID="9fc77a68-5c6e-4325-9846-42b1e4f286ea" containerID="f08a74b68b3512eab0caf35f07e1f1e08a7a11fe07dc70518c49d4eb79cd8f20" exitCode=0 Sep 29 10:41:02 crc kubenswrapper[4727]: I0929 10:41:02.174438 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh4r9" event={"ID":"9fc77a68-5c6e-4325-9846-42b1e4f286ea","Type":"ContainerDied","Data":"f08a74b68b3512eab0caf35f07e1f1e08a7a11fe07dc70518c49d4eb79cd8f20"} Sep 29 10:41:02 crc kubenswrapper[4727]: I0929 10:41:02.180486 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"fe7ab9d07bc6c16ef7d89c62656fe70c31f1aae5c3efd472969fe2265a0eea20"} Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.256778 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-sg2mp"] Sep 29 10:41:03 crc kubenswrapper[4727]: E0929 10:41:03.257617 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9030b15b-1cf7-4797-ae8e-b6b0a6137b92" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.257637 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9030b15b-1cf7-4797-ae8e-b6b0a6137b92" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: E0929 10:41:03.257664 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1ec441-5723-4541-8614-435ec4c53b07" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.257672 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1ec441-5723-4541-8614-435ec4c53b07" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.257885 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9030b15b-1cf7-4797-ae8e-b6b0a6137b92" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.257907 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1ec441-5723-4541-8614-435ec4c53b07" containerName="mariadb-account-create" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.258625 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.261079 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7bgkq" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.261289 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.265565 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sg2mp"] Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.278548 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.342746 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.342817 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.342846 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.342881 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.342930 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.343014 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwjr\" (UniqueName: \"kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.396870 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450283 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450382 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450415 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450454 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450475 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450526 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwjr\" (UniqueName: \"kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.450662 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.456882 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.458643 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.460134 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.463779 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.464826 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-8n57q"] Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.466093 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.477709 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.477885 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-45t5g" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.477999 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.491200 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwjr\" (UniqueName: \"kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr\") pod \"cinder-db-sync-sg2mp\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.495827 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8n57q"] Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.558630 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.558714 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzl8\" (UniqueName: \"kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.558780 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.583421 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.661721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.662595 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzl8\" (UniqueName: \"kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.662736 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.668995 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.669085 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.678775 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzl8\" (UniqueName: \"kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8\") pod \"neutron-db-sync-8n57q\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.878347 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.988077 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:41:03 crc kubenswrapper[4727]: I0929 10:41:03.995791 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cx7d2" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072006 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072424 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts\") pod \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072533 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs\") pod \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072619 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072710 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle\") pod \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072852 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072977 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.073046 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cswvr\" (UniqueName: \"kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.073187 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data\") pod \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.073330 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys\") pod \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\" (UID: \"9fc77a68-5c6e-4325-9846-42b1e4f286ea\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.073473 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tkfz\" (UniqueName: \"kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz\") pod \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\" (UID: \"f16d06b5-d52a-48bb-8809-f06e0959ea6e\") " Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.072968 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs" (OuterVolumeSpecName: "logs") pod "f16d06b5-d52a-48bb-8809-f06e0959ea6e" (UID: "f16d06b5-d52a-48bb-8809-f06e0959ea6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.089634 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts" (OuterVolumeSpecName: "scripts") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.089763 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.090095 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr" (OuterVolumeSpecName: "kube-api-access-cswvr") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "kube-api-access-cswvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.095537 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts" (OuterVolumeSpecName: "scripts") pod "f16d06b5-d52a-48bb-8809-f06e0959ea6e" (UID: "f16d06b5-d52a-48bb-8809-f06e0959ea6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.105651 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz" (OuterVolumeSpecName: "kube-api-access-8tkfz") pod "f16d06b5-d52a-48bb-8809-f06e0959ea6e" (UID: "f16d06b5-d52a-48bb-8809-f06e0959ea6e"). InnerVolumeSpecName "kube-api-access-8tkfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.108750 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.142955 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data" (OuterVolumeSpecName: "config-data") pod "f16d06b5-d52a-48bb-8809-f06e0959ea6e" (UID: "f16d06b5-d52a-48bb-8809-f06e0959ea6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.153606 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.165097 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f16d06b5-d52a-48bb-8809-f06e0959ea6e" (UID: "f16d06b5-d52a-48bb-8809-f06e0959ea6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175127 4727 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175168 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175176 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f16d06b5-d52a-48bb-8809-f06e0959ea6e-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175185 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175193 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175202 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175210 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cswvr\" (UniqueName: \"kubernetes.io/projected/9fc77a68-5c6e-4325-9846-42b1e4f286ea-kube-api-access-cswvr\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175220 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16d06b5-d52a-48bb-8809-f06e0959ea6e-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175227 4727 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.175235 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tkfz\" (UniqueName: \"kubernetes.io/projected/f16d06b5-d52a-48bb-8809-f06e0959ea6e-kube-api-access-8tkfz\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.180514 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data" (OuterVolumeSpecName: "config-data") pod "9fc77a68-5c6e-4325-9846-42b1e4f286ea" (UID: "9fc77a68-5c6e-4325-9846-42b1e4f286ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.204776 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cx7d2" event={"ID":"f16d06b5-d52a-48bb-8809-f06e0959ea6e","Type":"ContainerDied","Data":"ceb803bf985e14d6f4b3ec771fd470023b0ddfe62cdf8230d39118d07ac4c194"} Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.204827 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb803bf985e14d6f4b3ec771fd470023b0ddfe62cdf8230d39118d07ac4c194" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.204890 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cx7d2" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.211735 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh4r9" event={"ID":"9fc77a68-5c6e-4325-9846-42b1e4f286ea","Type":"ContainerDied","Data":"8eb08975bcb6303a931ad027c1a315a5d142440712d4db1645bbf5450d793819"} Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.211782 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eb08975bcb6303a931ad027c1a315a5d142440712d4db1645bbf5450d793819" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.211878 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh4r9" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.277219 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fc77a68-5c6e-4325-9846-42b1e4f286ea-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.339161 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64f85d6496-qmbdm"] Sep 29 10:41:04 crc kubenswrapper[4727]: E0929 10:41:04.340966 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f16d06b5-d52a-48bb-8809-f06e0959ea6e" containerName="placement-db-sync" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.340998 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f16d06b5-d52a-48bb-8809-f06e0959ea6e" containerName="placement-db-sync" Sep 29 10:41:04 crc kubenswrapper[4727]: E0929 10:41:04.341016 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc77a68-5c6e-4325-9846-42b1e4f286ea" containerName="keystone-bootstrap" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.341022 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc77a68-5c6e-4325-9846-42b1e4f286ea" containerName="keystone-bootstrap" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.341260 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f16d06b5-d52a-48bb-8809-f06e0959ea6e" containerName="placement-db-sync" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.341277 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc77a68-5c6e-4325-9846-42b1e4f286ea" containerName="keystone-bootstrap" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.341923 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.344394 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.346610 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.346735 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.346867 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2g6jb" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.346990 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.347048 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.356060 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64f85d6496-qmbdm"] Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.479895 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-scripts\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.479973 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-public-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480046 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-credential-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480063 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-config-data\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480120 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-fernet-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480149 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-internal-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480198 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvccc\" (UniqueName: \"kubernetes.io/projected/9eae012b-b92e-44ae-8396-f7331f267787-kube-api-access-hvccc\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.480228 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-combined-ca-bundle\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582088 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-scripts\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582149 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-public-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582199 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-credential-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582217 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-config-data\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582235 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-fernet-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582287 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-internal-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582311 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvccc\" (UniqueName: \"kubernetes.io/projected/9eae012b-b92e-44ae-8396-f7331f267787-kube-api-access-hvccc\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.582352 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-combined-ca-bundle\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.588266 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-combined-ca-bundle\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.590666 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-scripts\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.594715 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-credential-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.602567 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-config-data\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.608483 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-public-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.613265 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvccc\" (UniqueName: \"kubernetes.io/projected/9eae012b-b92e-44ae-8396-f7331f267787-kube-api-access-hvccc\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.617820 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-internal-tls-certs\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.626201 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eae012b-b92e-44ae-8396-f7331f267787-fernet-keys\") pod \"keystone-64f85d6496-qmbdm\" (UID: \"9eae012b-b92e-44ae-8396-f7331f267787\") " pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:04 crc kubenswrapper[4727]: I0929 10:41:04.673158 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.117890 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c44b4b898-64gpw"] Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.119835 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.122531 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.122734 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pcthw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.123002 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.123174 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.125649 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.125984 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c44b4b898-64gpw"] Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193188 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xbh\" (UniqueName: \"kubernetes.io/projected/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-kube-api-access-s8xbh\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193242 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-combined-ca-bundle\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193277 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-logs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193332 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-scripts\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193403 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-internal-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193429 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-config-data\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.193470 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-public-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.294706 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xbh\" (UniqueName: \"kubernetes.io/projected/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-kube-api-access-s8xbh\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.294797 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-combined-ca-bundle\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.294833 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-logs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.294907 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-scripts\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.294988 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-internal-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.295013 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-config-data\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.295042 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-public-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.298566 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-logs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.299161 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-public-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.299943 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-combined-ca-bundle\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.300854 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-internal-tls-certs\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.309128 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-config-data\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.310615 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-scripts\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.324422 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xbh\" (UniqueName: \"kubernetes.io/projected/1b18f96d-e3b8-4758-8f71-b75a40cc9fbe-kube-api-access-s8xbh\") pod \"placement-c44b4b898-64gpw\" (UID: \"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe\") " pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:05 crc kubenswrapper[4727]: I0929 10:41:05.453725 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:06 crc kubenswrapper[4727]: I0929 10:41:06.782806 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:06 crc kubenswrapper[4727]: I0929 10:41:06.783129 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:06 crc kubenswrapper[4727]: I0929 10:41:06.819693 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:06 crc kubenswrapper[4727]: I0929 10:41:06.831193 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.091385 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.091937 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.127804 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.130743 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.244159 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.244733 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.244752 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 10:41:07 crc kubenswrapper[4727]: I0929 10:41:07.244764 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.274168 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.275889 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.311530 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.311669 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.314924 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.480163 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.485267 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.664259 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sg2mp"] Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.742210 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64f85d6496-qmbdm"] Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.875476 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Sep 29 10:41:09 crc kubenswrapper[4727]: I0929 10:41:09.963040 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-54cbdfc6b8-xl7cd" podUID="4b7b8921-ae97-45b0-a082-07db49561514" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.242918 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c44b4b898-64gpw"] Sep 29 10:41:10 crc kubenswrapper[4727]: W0929 10:41:10.262677 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b18f96d_e3b8_4758_8f71_b75a40cc9fbe.slice/crio-93806576c9ed682fb90ecbe62d249abfccfb90ecc5258a683cd8abfbb46ef809 WatchSource:0}: Error finding container 93806576c9ed682fb90ecbe62d249abfccfb90ecc5258a683cd8abfbb46ef809: Status 404 returned error can't find the container with id 93806576c9ed682fb90ecbe62d249abfccfb90ecc5258a683cd8abfbb46ef809 Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.296501 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64f85d6496-qmbdm" event={"ID":"9eae012b-b92e-44ae-8396-f7331f267787","Type":"ContainerStarted","Data":"97c190dfe1f16d72e337cf31e22d1c216c783df812659a44794c9df9178d2cd2"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.296552 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64f85d6496-qmbdm" event={"ID":"9eae012b-b92e-44ae-8396-f7331f267787","Type":"ContainerStarted","Data":"4e2222edf5148afa31fff3f5b4d034e9aaeb408701e47e2e6e7378382596bfca"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.297815 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:10 crc kubenswrapper[4727]: W0929 10:41:10.341872 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ab43089_e386_4585_9324_79b4750eeeff.slice/crio-161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea WatchSource:0}: Error finding container 161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea: Status 404 returned error can't find the container with id 161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.342797 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8n57q"] Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.348443 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64f85d6496-qmbdm" podStartSLOduration=6.3484153 podStartE2EDuration="6.3484153s" podCreationTimestamp="2025-09-29 10:41:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:10.33095757 +0000 UTC m=+1140.504270952" watchObservedRunningTime="2025-09-29 10:41:10.3484153 +0000 UTC m=+1140.521728682" Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.365724 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"0cea4e9092591bc651086be97158a7f286687cd48ffef005102d1162a0ca5237"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.391081 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-68lgd" event={"ID":"74fb3935-9e35-4a18-a2e4-b4442581bc00","Type":"ContainerStarted","Data":"485e5cf3b6626ea3f28dc564f4341b93555e74bd4bb70dfc6245139ad307c02d"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.403839 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerStarted","Data":"5ebfc26ae2a9cf2c46c435148a9ff0c700f942f1cfe1cc9d677a14c7c521b5f2"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.414464 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44b4b898-64gpw" event={"ID":"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe","Type":"ContainerStarted","Data":"93806576c9ed682fb90ecbe62d249abfccfb90ecc5258a683cd8abfbb46ef809"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.416746 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sg2mp" event={"ID":"62cb5972-95c1-4963-b0f2-bfee701d6453","Type":"ContainerStarted","Data":"0e37e5709f7ca6384260459d6dfd3c5cff249d90a8d796ac047c19791082e8b2"} Sep 29 10:41:10 crc kubenswrapper[4727]: I0929 10:41:10.417200 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-68lgd" podStartSLOduration=1.480279717 podStartE2EDuration="12.417178829s" podCreationTimestamp="2025-09-29 10:40:58 +0000 UTC" firstStartedPulling="2025-09-29 10:40:58.896857906 +0000 UTC m=+1129.070171268" lastFinishedPulling="2025-09-29 10:41:09.833757018 +0000 UTC m=+1140.007070380" observedRunningTime="2025-09-29 10:41:10.411354339 +0000 UTC m=+1140.584667701" watchObservedRunningTime="2025-09-29 10:41:10.417178829 +0000 UTC m=+1140.590492191" Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.437399 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44b4b898-64gpw" event={"ID":"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe","Type":"ContainerStarted","Data":"cd9a188670f17607d8164faffc06e23681ba58998c72ff447fded8834c59194b"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.438269 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44b4b898-64gpw" event={"ID":"1b18f96d-e3b8-4758-8f71-b75a40cc9fbe","Type":"ContainerStarted","Data":"cf601d73db30a22168e8a56984b6c0dadde306b3b6f65b3f93709341319b015e"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.438413 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.438441 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.460051 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8n57q" event={"ID":"2ab43089-e386-4585-9324-79b4750eeeff","Type":"ContainerStarted","Data":"a05bb2c1ba81f073e36943b5bbad927ad3c500e9367e937bf7845db8d01ad98a"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.460122 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8n57q" event={"ID":"2ab43089-e386-4585-9324-79b4750eeeff","Type":"ContainerStarted","Data":"161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.511314 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c44b4b898-64gpw" podStartSLOduration=6.511289248 podStartE2EDuration="6.511289248s" podCreationTimestamp="2025-09-29 10:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:11.503077377 +0000 UTC m=+1141.676390749" watchObservedRunningTime="2025-09-29 10:41:11.511289248 +0000 UTC m=+1141.684602600" Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.522587 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"b9965e8e4fbe2a7adb6566fae1d5f943512704c26d1f19fe4248c6b49bfd0265"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.522625 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"d1308bce06da586619f2cdef2fec885818209260e39e52b7d19477736194fc18"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.522636 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"cc78d80e57dce2f36d65df5d87b9ea3edf6d38b5ded452586a3bb432d235d124"} Sep 29 10:41:11 crc kubenswrapper[4727]: I0929 10:41:11.530679 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-8n57q" podStartSLOduration=8.530663427 podStartE2EDuration="8.530663427s" podCreationTimestamp="2025-09-29 10:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:11.524170919 +0000 UTC m=+1141.697484281" watchObservedRunningTime="2025-09-29 10:41:11.530663427 +0000 UTC m=+1141.703976789" Sep 29 10:41:12 crc kubenswrapper[4727]: I0929 10:41:12.539107 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"d76553e680eab567aa54762fdab8ca8553f75fb6731ff981f928d174cddce4ab"} Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.575807 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"7ce0fbc447c9a3133cfb693671e3e34bb36d627613b83138a9792e0963299507"} Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.576082 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7a6a9dee-ca84-4c2d-abc9-613bccd90764","Type":"ContainerStarted","Data":"da56ffe00e9b3289163f28dd9f8b9950c7341d242c6a17bec48268395c7e7669"} Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.614445 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=70.056025881 podStartE2EDuration="1m37.61442603s" podCreationTimestamp="2025-09-29 10:39:36 +0000 UTC" firstStartedPulling="2025-09-29 10:40:42.286016373 +0000 UTC m=+1112.459329735" lastFinishedPulling="2025-09-29 10:41:09.844416522 +0000 UTC m=+1140.017729884" observedRunningTime="2025-09-29 10:41:13.614204864 +0000 UTC m=+1143.787518246" watchObservedRunningTime="2025-09-29 10:41:13.61442603 +0000 UTC m=+1143.787739392" Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.923212 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.929353 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.934284 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:13 crc kubenswrapper[4727]: I0929 10:41:13.944226 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006071 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006133 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006214 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006240 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006267 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n68nj\" (UniqueName: \"kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.006298 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107641 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107708 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n68nj\" (UniqueName: \"kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107758 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107817 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107863 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.107955 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.108725 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.108796 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.109330 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.109482 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.110250 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.135369 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n68nj\" (UniqueName: \"kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj\") pod \"dnsmasq-dns-84d74cbc87-92jsg\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.248435 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.588779 4727 generic.go:334] "Generic (PLEG): container finished" podID="74fb3935-9e35-4a18-a2e4-b4442581bc00" containerID="485e5cf3b6626ea3f28dc564f4341b93555e74bd4bb70dfc6245139ad307c02d" exitCode=0 Sep 29 10:41:14 crc kubenswrapper[4727]: I0929 10:41:14.589447 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-68lgd" event={"ID":"74fb3935-9e35-4a18-a2e4-b4442581bc00","Type":"ContainerDied","Data":"485e5cf3b6626ea3f28dc564f4341b93555e74bd4bb70dfc6245139ad307c02d"} Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.356754 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-68lgd" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.495269 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle\") pod \"74fb3935-9e35-4a18-a2e4-b4442581bc00\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.495397 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data\") pod \"74fb3935-9e35-4a18-a2e4-b4442581bc00\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.495455 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc9pq\" (UniqueName: \"kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq\") pod \"74fb3935-9e35-4a18-a2e4-b4442581bc00\" (UID: \"74fb3935-9e35-4a18-a2e4-b4442581bc00\") " Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.513948 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "74fb3935-9e35-4a18-a2e4-b4442581bc00" (UID: "74fb3935-9e35-4a18-a2e4-b4442581bc00"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.514006 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq" (OuterVolumeSpecName: "kube-api-access-rc9pq") pod "74fb3935-9e35-4a18-a2e4-b4442581bc00" (UID: "74fb3935-9e35-4a18-a2e4-b4442581bc00"). InnerVolumeSpecName "kube-api-access-rc9pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.526426 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74fb3935-9e35-4a18-a2e4-b4442581bc00" (UID: "74fb3935-9e35-4a18-a2e4-b4442581bc00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.597951 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc9pq\" (UniqueName: \"kubernetes.io/projected/74fb3935-9e35-4a18-a2e4-b4442581bc00-kube-api-access-rc9pq\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.597982 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.597992 4727 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/74fb3935-9e35-4a18-a2e4-b4442581bc00-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.638833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-68lgd" event={"ID":"74fb3935-9e35-4a18-a2e4-b4442581bc00","Type":"ContainerDied","Data":"b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554"} Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.638884 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b14cdfdc2bc419e8a56f7eed8ffdbcd7f12a7b1fb6ea40e158e113b23d0ae554" Sep 29 10:41:18 crc kubenswrapper[4727]: I0929 10:41:18.638889 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-68lgd" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.266749 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.266797 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.722509 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-fb4b7877c-kksr7"] Sep 29 10:41:19 crc kubenswrapper[4727]: E0929 10:41:19.722886 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fb3935-9e35-4a18-a2e4-b4442581bc00" containerName="barbican-db-sync" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.722897 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fb3935-9e35-4a18-a2e4-b4442581bc00" containerName="barbican-db-sync" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.723073 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fb3935-9e35-4a18-a2e4-b4442581bc00" containerName="barbican-db-sync" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.723988 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.727776 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.728198 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-wx6p9" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.728427 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.772662 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb4b7877c-kksr7"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.800228 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.800283 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-combined-ca-bundle\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.801981 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d320a-33b6-4df1-bf0e-da396f28f90e-logs\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.804979 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbh6c\" (UniqueName: \"kubernetes.io/projected/134d320a-33b6-4df1-bf0e-da396f28f90e-kube-api-access-kbh6c\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.805163 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data-custom\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.809131 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.846990 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.849428 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.861479 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.877167 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.897873 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-58b58bf9d-2f9w7"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.899424 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.903137 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906301 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d320a-33b6-4df1-bf0e-da396f28f90e-logs\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906394 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbh6c\" (UniqueName: \"kubernetes.io/projected/134d320a-33b6-4df1-bf0e-da396f28f90e-kube-api-access-kbh6c\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906422 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcmnq\" (UniqueName: \"kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906441 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906463 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906504 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data-custom\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906523 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906571 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906598 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-combined-ca-bundle\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906618 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.906649 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.907476 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d320a-33b6-4df1-bf0e-da396f28f90e-logs\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.918180 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data-custom\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.922959 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-combined-ca-bundle\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.924479 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b58bf9d-2f9w7"] Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.939681 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbh6c\" (UniqueName: \"kubernetes.io/projected/134d320a-33b6-4df1-bf0e-da396f28f90e-kube-api-access-kbh6c\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.941291 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134d320a-33b6-4df1-bf0e-da396f28f90e-config-data\") pod \"barbican-worker-fb4b7877c-kksr7\" (UID: \"134d320a-33b6-4df1-bf0e-da396f28f90e\") " pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:19 crc kubenswrapper[4727]: I0929 10:41:19.967575 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-54cbdfc6b8-xl7cd" podUID="4b7b8921-ae97-45b0-a082-07db49561514" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008136 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008347 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008500 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008542 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data-custom\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008589 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978ca6db-0819-4482-a407-a11ea6ebc3f2-logs\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008675 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87wxn\" (UniqueName: \"kubernetes.io/projected/978ca6db-0819-4482-a407-a11ea6ebc3f2-kube-api-access-87wxn\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008726 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-combined-ca-bundle\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008753 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcmnq\" (UniqueName: \"kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008779 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.008832 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.010162 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.010782 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.010928 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.011080 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.011540 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.021532 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.023540 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.035576 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.039360 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcmnq\" (UniqueName: \"kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq\") pod \"dnsmasq-dns-85cb9df8d7-5dc48\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.041213 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.056622 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb4b7877c-kksr7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113751 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113834 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data-custom\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113855 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978ca6db-0819-4482-a407-a11ea6ebc3f2-logs\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113903 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87wxn\" (UniqueName: \"kubernetes.io/projected/978ca6db-0819-4482-a407-a11ea6ebc3f2-kube-api-access-87wxn\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113926 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113946 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113963 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-combined-ca-bundle\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.113990 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.114007 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.114023 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tspq4\" (UniqueName: \"kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.114826 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978ca6db-0819-4482-a407-a11ea6ebc3f2-logs\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.118903 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data-custom\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.129616 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-config-data\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.133613 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ca6db-0819-4482-a407-a11ea6ebc3f2-combined-ca-bundle\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.134671 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87wxn\" (UniqueName: \"kubernetes.io/projected/978ca6db-0819-4482-a407-a11ea6ebc3f2-kube-api-access-87wxn\") pod \"barbican-keystone-listener-58b58bf9d-2f9w7\" (UID: \"978ca6db-0819-4482-a407-a11ea6ebc3f2\") " pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.177007 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.216592 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.216666 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.216698 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.216722 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.216743 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tspq4\" (UniqueName: \"kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.219605 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.222097 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.222389 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.225490 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.239918 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tspq4\" (UniqueName: \"kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4\") pod \"barbican-api-64b84775bb-fqb4p\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.304428 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" Sep 29 10:41:20 crc kubenswrapper[4727]: I0929 10:41:20.392456 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.349075 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bc65c799d-pbjqp"] Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.351352 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.354326 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.354326 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.374891 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc65c799d-pbjqp"] Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496419 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-internal-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496491 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8k4b\" (UniqueName: \"kubernetes.io/projected/7529e49e-8904-4b11-ba31-ed805fc0fee2-kube-api-access-j8k4b\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496535 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-public-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496691 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data-custom\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496720 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7529e49e-8904-4b11-ba31-ed805fc0fee2-logs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496764 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.496895 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-combined-ca-bundle\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599295 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data-custom\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599376 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7529e49e-8904-4b11-ba31-ed805fc0fee2-logs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599423 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599463 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-combined-ca-bundle\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599539 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-internal-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599563 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8k4b\" (UniqueName: \"kubernetes.io/projected/7529e49e-8904-4b11-ba31-ed805fc0fee2-kube-api-access-j8k4b\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599601 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-public-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.599936 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7529e49e-8904-4b11-ba31-ed805fc0fee2-logs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.604979 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-public-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.605816 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data-custom\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.606290 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-config-data\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.606724 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-internal-tls-certs\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.612058 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7529e49e-8904-4b11-ba31-ed805fc0fee2-combined-ca-bundle\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.617760 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8k4b\" (UniqueName: \"kubernetes.io/projected/7529e49e-8904-4b11-ba31-ed805fc0fee2-kube-api-access-j8k4b\") pod \"barbican-api-7bc65c799d-pbjqp\" (UID: \"7529e49e-8904-4b11-ba31-ed805fc0fee2\") " pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:22 crc kubenswrapper[4727]: I0929 10:41:22.677378 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:27 crc kubenswrapper[4727]: I0929 10:41:27.724676 4727 generic.go:334] "Generic (PLEG): container finished" podID="b0b648b7-c07b-4516-9365-3b7680519463" containerID="6eb109f03a51bf37afc95e6079b2f74e3158c6e40ed77d59f590d891d0ea76c0" exitCode=137 Sep 29 10:41:27 crc kubenswrapper[4727]: I0929 10:41:27.725307 4727 generic.go:334] "Generic (PLEG): container finished" podID="b0b648b7-c07b-4516-9365-3b7680519463" containerID="fa3b74a407f12888522d7e2dca80344ae4eb404a0123f5b27ed343722dda4ddd" exitCode=137 Sep 29 10:41:27 crc kubenswrapper[4727]: I0929 10:41:27.724863 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerDied","Data":"6eb109f03a51bf37afc95e6079b2f74e3158c6e40ed77d59f590d891d0ea76c0"} Sep 29 10:41:27 crc kubenswrapper[4727]: I0929 10:41:27.725377 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerDied","Data":"fa3b74a407f12888522d7e2dca80344ae4eb404a0123f5b27ed343722dda4ddd"} Sep 29 10:41:29 crc kubenswrapper[4727]: E0929 10:41:29.762760 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Sep 29 10:41:29 crc kubenswrapper[4727]: E0929 10:41:29.763327 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nphtd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(12d4eb15-f153-4824-85a4-40bceb032c7f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 29 10:41:29 crc kubenswrapper[4727]: E0929 10:41:29.764594 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" Sep 29 10:41:30 crc kubenswrapper[4727]: I0929 10:41:30.751674 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="ceilometer-notification-agent" containerID="cri-o://4159dc13afebbe60f4ba315b81a3748e8cc5deb4542e1bec8aa38974b395c8a8" gracePeriod=30 Sep 29 10:41:30 crc kubenswrapper[4727]: I0929 10:41:30.751851 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="sg-core" containerID="cri-o://5ebfc26ae2a9cf2c46c435148a9ff0c700f942f1cfe1cc9d677a14c7c521b5f2" gracePeriod=30 Sep 29 10:41:31 crc kubenswrapper[4727]: E0929 10:41:31.026598 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Sep 29 10:41:31 crc kubenswrapper[4727]: E0929 10:41:31.027127 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgwjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-sg2mp_openstack(62cb5972-95c1-4963-b0f2-bfee701d6453): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 10:41:31 crc kubenswrapper[4727]: E0929 10:41:31.028465 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-sg2mp" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.370419 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.490704 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key\") pod \"b0b648b7-c07b-4516-9365-3b7680519463\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.490913 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs\") pod \"b0b648b7-c07b-4516-9365-3b7680519463\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.490978 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtcz2\" (UniqueName: \"kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2\") pod \"b0b648b7-c07b-4516-9365-3b7680519463\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.491036 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts\") pod \"b0b648b7-c07b-4516-9365-3b7680519463\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.491110 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data\") pod \"b0b648b7-c07b-4516-9365-3b7680519463\" (UID: \"b0b648b7-c07b-4516-9365-3b7680519463\") " Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.492696 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs" (OuterVolumeSpecName: "logs") pod "b0b648b7-c07b-4516-9365-3b7680519463" (UID: "b0b648b7-c07b-4516-9365-3b7680519463"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.497778 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2" (OuterVolumeSpecName: "kube-api-access-vtcz2") pod "b0b648b7-c07b-4516-9365-3b7680519463" (UID: "b0b648b7-c07b-4516-9365-3b7680519463"). InnerVolumeSpecName "kube-api-access-vtcz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.497984 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b0b648b7-c07b-4516-9365-3b7680519463" (UID: "b0b648b7-c07b-4516-9365-3b7680519463"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.520592 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data" (OuterVolumeSpecName: "config-data") pod "b0b648b7-c07b-4516-9365-3b7680519463" (UID: "b0b648b7-c07b-4516-9365-3b7680519463"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.524400 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts" (OuterVolumeSpecName: "scripts") pod "b0b648b7-c07b-4516-9365-3b7680519463" (UID: "b0b648b7-c07b-4516-9365-3b7680519463"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.593726 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0b648b7-c07b-4516-9365-3b7680519463-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.593777 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtcz2\" (UniqueName: \"kubernetes.io/projected/b0b648b7-c07b-4516-9365-3b7680519463-kube-api-access-vtcz2\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.593791 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.593800 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0b648b7-c07b-4516-9365-3b7680519463-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.593809 4727 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0b648b7-c07b-4516-9365-3b7680519463-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.721148 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b58bf9d-2f9w7"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.733884 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:31 crc kubenswrapper[4727]: W0929 10:41:31.737954 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978ca6db_0819_4482_a407_a11ea6ebc3f2.slice/crio-f88ecad1fdcdf71cc508d405f4249af23b75ff0501fa5c7307ffcb66739759eb WatchSource:0}: Error finding container f88ecad1fdcdf71cc508d405f4249af23b75ff0501fa5c7307ffcb66739759eb: Status 404 returned error can't find the container with id f88ecad1fdcdf71cc508d405f4249af23b75ff0501fa5c7307ffcb66739759eb Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.765575 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65fc4fcb97-6gnq7" event={"ID":"b0b648b7-c07b-4516-9365-3b7680519463","Type":"ContainerDied","Data":"8c53b925fe4c11ede16977b8599099dbf049e124e6f04c5daa60e792563a9afd"} Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.765610 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65fc4fcb97-6gnq7" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.765909 4727 scope.go:117] "RemoveContainer" containerID="6eb109f03a51bf37afc95e6079b2f74e3158c6e40ed77d59f590d891d0ea76c0" Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.780845 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" event={"ID":"71eeedc3-6697-44a3-9953-71f445762b90","Type":"ContainerStarted","Data":"f295f923c935b83c7f1938e14cb510c02fc5ebfc393adee3f7f166e37e2ddc19"} Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.784419 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" event={"ID":"978ca6db-0819-4482-a407-a11ea6ebc3f2","Type":"ContainerStarted","Data":"f88ecad1fdcdf71cc508d405f4249af23b75ff0501fa5c7307ffcb66739759eb"} Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.819474 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb4b7877c-kksr7"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.827043 4727 generic.go:334] "Generic (PLEG): container finished" podID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerID="5ebfc26ae2a9cf2c46c435148a9ff0c700f942f1cfe1cc9d677a14c7c521b5f2" exitCode=2 Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.827142 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerDied","Data":"5ebfc26ae2a9cf2c46c435148a9ff0c700f942f1cfe1cc9d677a14c7c521b5f2"} Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.844053 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc65c799d-pbjqp"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.851778 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.860996 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.870453 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.878881 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65fc4fcb97-6gnq7"] Sep 29 10:41:31 crc kubenswrapper[4727]: I0929 10:41:31.961511 4727 scope.go:117] "RemoveContainer" containerID="fa3b74a407f12888522d7e2dca80344ae4eb404a0123f5b27ed343722dda4ddd" Sep 29 10:41:31 crc kubenswrapper[4727]: E0929 10:41:31.961558 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-sg2mp" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" Sep 29 10:41:31 crc kubenswrapper[4727]: W0929 10:41:31.965705 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd673ffa_5d52_490b_8065_36767a0b53cf.slice/crio-3182d4cbfa73fc7ce7c154132f3bc54cdebde1142de39b85534186d9eaf19553 WatchSource:0}: Error finding container 3182d4cbfa73fc7ce7c154132f3bc54cdebde1142de39b85534186d9eaf19553: Status 404 returned error can't find the container with id 3182d4cbfa73fc7ce7c154132f3bc54cdebde1142de39b85534186d9eaf19553 Sep 29 10:41:31 crc kubenswrapper[4727]: W0929 10:41:31.973231 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ac2ffa4_3d80_4119_abb3_0a91ed4e3ba8.slice/crio-c12faa2ead0ed9393b7fd110b003189f6acb3887b6f16a9e96c959cf2d9f9669 WatchSource:0}: Error finding container c12faa2ead0ed9393b7fd110b003189f6acb3887b6f16a9e96c959cf2d9f9669: Status 404 returned error can't find the container with id c12faa2ead0ed9393b7fd110b003189f6acb3887b6f16a9e96c959cf2d9f9669 Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.045601 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.060604 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.839616 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerStarted","Data":"0bfdacb4ed15db81f9496ff96bd31129b7c79e1e82a92bc6231ee0a95c6b0e68"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.839927 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerStarted","Data":"03b780a0d9369fd3b6cd419c58c239b1ece3bf21c04aa3f5f396da48aa07503f"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.839944 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerStarted","Data":"3182d4cbfa73fc7ce7c154132f3bc54cdebde1142de39b85534186d9eaf19553"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.840262 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.840284 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.842629 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb4b7877c-kksr7" event={"ID":"134d320a-33b6-4df1-bf0e-da396f28f90e","Type":"ContainerStarted","Data":"51b4a712b780abc480f2251203cb4aaf9bcd94abe8f64a809e2cd722b0bc31f9"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.844970 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc65c799d-pbjqp" event={"ID":"7529e49e-8904-4b11-ba31-ed805fc0fee2","Type":"ContainerStarted","Data":"58028b2d2a7dcad8b4c33f88cee83e20fe4b9213d94d4165475e2b5c952f3d55"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.845008 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc65c799d-pbjqp" event={"ID":"7529e49e-8904-4b11-ba31-ed805fc0fee2","Type":"ContainerStarted","Data":"7b7b94119d2e74de01e1d377aee3241c7619490e0dc2226cc790cca7cd5661b4"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.845020 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc65c799d-pbjqp" event={"ID":"7529e49e-8904-4b11-ba31-ed805fc0fee2","Type":"ContainerStarted","Data":"00d94284fd94201d4bbb97b8ca620cd2c3cba742c3de8764c16b85a2af9e4efc"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.845545 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.845572 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.848400 4727 generic.go:334] "Generic (PLEG): container finished" podID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerID="0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993" exitCode=0 Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.848453 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" event={"ID":"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8","Type":"ContainerDied","Data":"0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.848472 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" event={"ID":"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8","Type":"ContainerStarted","Data":"c12faa2ead0ed9393b7fd110b003189f6acb3887b6f16a9e96c959cf2d9f9669"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.851662 4727 generic.go:334] "Generic (PLEG): container finished" podID="71eeedc3-6697-44a3-9953-71f445762b90" containerID="e0d4806cd702207b60db343cb2767fe32f103d7ea303f502b09b751dc3f58525" exitCode=0 Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.851688 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" event={"ID":"71eeedc3-6697-44a3-9953-71f445762b90","Type":"ContainerDied","Data":"e0d4806cd702207b60db343cb2767fe32f103d7ea303f502b09b751dc3f58525"} Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.857831 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-64b84775bb-fqb4p" podStartSLOduration=13.857813845 podStartE2EDuration="13.857813845s" podCreationTimestamp="2025-09-29 10:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:32.85760702 +0000 UTC m=+1163.030920402" watchObservedRunningTime="2025-09-29 10:41:32.857813845 +0000 UTC m=+1163.031127207" Sep 29 10:41:32 crc kubenswrapper[4727]: I0929 10:41:32.918179 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bc65c799d-pbjqp" podStartSLOduration=10.918146297 podStartE2EDuration="10.918146297s" podCreationTimestamp="2025-09-29 10:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:32.904957258 +0000 UTC m=+1163.078270620" watchObservedRunningTime="2025-09-29 10:41:32.918146297 +0000 UTC m=+1163.091459659" Sep 29 10:41:33 crc kubenswrapper[4727]: I0929 10:41:33.207530 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0b648b7-c07b-4516-9365-3b7680519463" path="/var/lib/kubelet/pods/b0b648b7-c07b-4516-9365-3b7680519463/volumes" Sep 29 10:41:33 crc kubenswrapper[4727]: I0929 10:41:33.860977 4727 generic.go:334] "Generic (PLEG): container finished" podID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerID="4159dc13afebbe60f4ba315b81a3748e8cc5deb4542e1bec8aa38974b395c8a8" exitCode=0 Sep 29 10:41:33 crc kubenswrapper[4727]: I0929 10:41:33.861358 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerDied","Data":"4159dc13afebbe60f4ba315b81a3748e8cc5deb4542e1bec8aa38974b395c8a8"} Sep 29 10:41:33 crc kubenswrapper[4727]: I0929 10:41:33.982029 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:41:33 crc kubenswrapper[4727]: I0929 10:41:33.987216 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069460 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069594 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069638 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069699 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n68nj\" (UniqueName: \"kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069735 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.069826 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb\") pod \"71eeedc3-6697-44a3-9953-71f445762b90\" (UID: \"71eeedc3-6697-44a3-9953-71f445762b90\") " Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.090518 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj" (OuterVolumeSpecName: "kube-api-access-n68nj") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "kube-api-access-n68nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.126488 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.126813 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.140414 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.144865 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.152496 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config" (OuterVolumeSpecName: "config") pod "71eeedc3-6697-44a3-9953-71f445762b90" (UID: "71eeedc3-6697-44a3-9953-71f445762b90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172678 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172723 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172744 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172760 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172777 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n68nj\" (UniqueName: \"kubernetes.io/projected/71eeedc3-6697-44a3-9953-71f445762b90-kube-api-access-n68nj\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.172795 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71eeedc3-6697-44a3-9953-71f445762b90-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.180805 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-54cbdfc6b8-xl7cd" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.241591 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.870376 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" event={"ID":"71eeedc3-6697-44a3-9953-71f445762b90","Type":"ContainerDied","Data":"f295f923c935b83c7f1938e14cb510c02fc5ebfc393adee3f7f166e37e2ddc19"} Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.870816 4727 scope.go:117] "RemoveContainer" containerID="e0d4806cd702207b60db343cb2767fe32f103d7ea303f502b09b751dc3f58525" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.870462 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d74cbc87-92jsg" Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.870632 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" containerID="cri-o://d4e2464b9ae71a314ef444db324581c200bec588ddbb7b07b1c93c3d46fa4633" gracePeriod=30 Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.870558 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon-log" containerID="cri-o://4cd6244f38b2627ab0c262de29eecfcff1b39f6d175990a0f329ccd2991f7f30" gracePeriod=30 Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.941257 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:34 crc kubenswrapper[4727]: I0929 10:41:34.948253 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84d74cbc87-92jsg"] Sep 29 10:41:35 crc kubenswrapper[4727]: I0929 10:41:35.119373 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71eeedc3-6697-44a3-9953-71f445762b90" path="/var/lib/kubelet/pods/71eeedc3-6697-44a3-9953-71f445762b90/volumes" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.146248 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242063 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242117 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242160 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nphtd\" (UniqueName: \"kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242223 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242271 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242323 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.242399 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml\") pod \"12d4eb15-f153-4824-85a4-40bceb032c7f\" (UID: \"12d4eb15-f153-4824-85a4-40bceb032c7f\") " Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.244313 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.244644 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.251188 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts" (OuterVolumeSpecName: "scripts") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.259526 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd" (OuterVolumeSpecName: "kube-api-access-nphtd") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "kube-api-access-nphtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.282183 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.293973 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data" (OuterVolumeSpecName: "config-data") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.299878 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "12d4eb15-f153-4824-85a4-40bceb032c7f" (UID: "12d4eb15-f153-4824-85a4-40bceb032c7f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.317948 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64f85d6496-qmbdm" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.350995 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351042 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351056 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351069 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351081 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nphtd\" (UniqueName: \"kubernetes.io/projected/12d4eb15-f153-4824-85a4-40bceb032c7f-kube-api-access-nphtd\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351092 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d4eb15-f153-4824-85a4-40bceb032c7f-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.351102 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12d4eb15-f153-4824-85a4-40bceb032c7f-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.907193 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" event={"ID":"978ca6db-0819-4482-a407-a11ea6ebc3f2","Type":"ContainerStarted","Data":"421d08c4b8bf9c274bef68a9185ce918167d46b80e9cbc5124ecc4717c384cc3"} Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.908485 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb4b7877c-kksr7" event={"ID":"134d320a-33b6-4df1-bf0e-da396f28f90e","Type":"ContainerStarted","Data":"c4c7a0dd93f95f21324c6eccabf7db57306b4494a0d28a2ec1b3fa728b69fa82"} Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.910362 4727 generic.go:334] "Generic (PLEG): container finished" podID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerID="d4e2464b9ae71a314ef444db324581c200bec588ddbb7b07b1c93c3d46fa4633" exitCode=0 Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.910419 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerDied","Data":"d4e2464b9ae71a314ef444db324581c200bec588ddbb7b07b1c93c3d46fa4633"} Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.912179 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12d4eb15-f153-4824-85a4-40bceb032c7f","Type":"ContainerDied","Data":"414ca8c023e12a35c9e960d6a63fb48b1432262aeebc2e928a35cba503d742dd"} Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.912212 4727 scope.go:117] "RemoveContainer" containerID="5ebfc26ae2a9cf2c46c435148a9ff0c700f942f1cfe1cc9d677a14c7c521b5f2" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.912233 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.914243 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" event={"ID":"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8","Type":"ContainerStarted","Data":"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c"} Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.914414 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.931594 4727 scope.go:117] "RemoveContainer" containerID="4159dc13afebbe60f4ba315b81a3748e8cc5deb4542e1bec8aa38974b395c8a8" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.942017 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" podStartSLOduration=19.942002594 podStartE2EDuration="19.942002594s" podCreationTimestamp="2025-09-29 10:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:38.941145301 +0000 UTC m=+1169.114458663" watchObservedRunningTime="2025-09-29 10:41:38.942002594 +0000 UTC m=+1169.115315956" Sep 29 10:41:38 crc kubenswrapper[4727]: I0929 10:41:38.987520 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.001511 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.020443 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:39 crc kubenswrapper[4727]: E0929 10:41:39.020927 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71eeedc3-6697-44a3-9953-71f445762b90" containerName="init" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.020951 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="71eeedc3-6697-44a3-9953-71f445762b90" containerName="init" Sep 29 10:41:39 crc kubenswrapper[4727]: E0929 10:41:39.020982 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon-log" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.020991 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon-log" Sep 29 10:41:39 crc kubenswrapper[4727]: E0929 10:41:39.021010 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="ceilometer-notification-agent" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021021 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="ceilometer-notification-agent" Sep 29 10:41:39 crc kubenswrapper[4727]: E0929 10:41:39.021052 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="sg-core" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021063 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="sg-core" Sep 29 10:41:39 crc kubenswrapper[4727]: E0929 10:41:39.021086 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021093 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021317 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="71eeedc3-6697-44a3-9953-71f445762b90" containerName="init" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021400 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021420 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="sg-core" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021433 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0b648b7-c07b-4516-9365-3b7680519463" containerName="horizon-log" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.021445 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" containerName="ceilometer-notification-agent" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.023021 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.030348 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.030819 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.046504 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.132690 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d4eb15-f153-4824-85a4-40bceb032c7f" path="/var/lib/kubelet/pods/12d4eb15-f153-4824-85a4-40bceb032c7f/volumes" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172457 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172524 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172613 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn24b\" (UniqueName: \"kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172750 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172827 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172857 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.172948 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274315 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274397 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn24b\" (UniqueName: \"kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274447 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274489 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274514 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274555 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274614 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274688 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.274884 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.280103 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.280435 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.283119 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.292897 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.310970 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn24b\" (UniqueName: \"kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b\") pod \"ceilometer-0\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.407541 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.409301 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.461248 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c44b4b898-64gpw" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.685781 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.862020 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.937773 4727 generic.go:334] "Generic (PLEG): container finished" podID="2ab43089-e386-4585-9324-79b4750eeeff" containerID="a05bb2c1ba81f073e36943b5bbad927ad3c500e9367e937bf7845db8d01ad98a" exitCode=0 Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.937848 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8n57q" event={"ID":"2ab43089-e386-4585-9324-79b4750eeeff","Type":"ContainerDied","Data":"a05bb2c1ba81f073e36943b5bbad927ad3c500e9367e937bf7845db8d01ad98a"} Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.957696 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" event={"ID":"978ca6db-0819-4482-a407-a11ea6ebc3f2","Type":"ContainerStarted","Data":"b0fc7cb9cc7f7fff9a3f3a2037ee364cefe990f51dde535691d23cbda2241572"} Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.969996 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb4b7877c-kksr7" event={"ID":"134d320a-33b6-4df1-bf0e-da396f28f90e","Type":"ContainerStarted","Data":"ae6531e2dbde72825f7cf6a631198d0ce2cc6bbe113433d5b309ae3a9bc3aaf2"} Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.984508 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:39 crc kubenswrapper[4727]: I0929 10:41:39.991677 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-58b58bf9d-2f9w7" podStartSLOduration=14.344583429 podStartE2EDuration="20.991662369s" podCreationTimestamp="2025-09-29 10:41:19 +0000 UTC" firstStartedPulling="2025-09-29 10:41:31.741187206 +0000 UTC m=+1161.914500568" lastFinishedPulling="2025-09-29 10:41:38.388266146 +0000 UTC m=+1168.561579508" observedRunningTime="2025-09-29 10:41:39.977133105 +0000 UTC m=+1170.150446467" watchObservedRunningTime="2025-09-29 10:41:39.991662369 +0000 UTC m=+1170.164975731" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.016707 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-fb4b7877c-kksr7" podStartSLOduration=14.587566881 podStartE2EDuration="21.016691633s" podCreationTimestamp="2025-09-29 10:41:19 +0000 UTC" firstStartedPulling="2025-09-29 10:41:31.993711024 +0000 UTC m=+1162.167024386" lastFinishedPulling="2025-09-29 10:41:38.422835776 +0000 UTC m=+1168.596149138" observedRunningTime="2025-09-29 10:41:40.014575659 +0000 UTC m=+1170.187889011" watchObservedRunningTime="2025-09-29 10:41:40.016691633 +0000 UTC m=+1170.190004995" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.432741 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.434271 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.438847 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-tr44t" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.438888 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.438847 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.481406 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.499307 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nz6m\" (UniqueName: \"kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.499364 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.499411 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.499489 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.504847 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bc65c799d-pbjqp" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.577889 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.578912 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" containerID="cri-o://03b780a0d9369fd3b6cd419c58c239b1ece3bf21c04aa3f5f396da48aa07503f" gracePeriod=30 Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.579289 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" containerID="cri-o://0bfdacb4ed15db81f9496ff96bd31129b7c79e1e82a92bc6231ee0a95c6b0e68" gracePeriod=30 Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.593749 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.594243 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.598648 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.598831 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.600895 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nz6m\" (UniqueName: \"kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.600931 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.600957 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.601056 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.601967 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.610546 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.659025 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nz6m\" (UniqueName: \"kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.664286 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.683715 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.686489 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.726413 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.764281 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.765576 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.783473 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.920227 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.920595 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config-secret\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.920703 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gmmb\" (UniqueName: \"kubernetes.io/projected/60736cbf-afd3-410e-87b8-bcf852ff2221-kube-api-access-2gmmb\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:40 crc kubenswrapper[4727]: I0929 10:41:40.920733 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-combined-ca-bundle\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: E0929 10:41:41.012656 4727 log.go:32] "RunPodSandbox from runtime service failed" err=< Sep 29 10:41:41 crc kubenswrapper[4727]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_2d102b65-bf8c-4df4-bca3-eadf749de206_0(821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92" Netns:"/var/run/netns/a4127118-3f52-4ab7-9bfe-0b8e4d45486b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92;K8S_POD_UID=2d102b65-bf8c-4df4-bca3-eadf749de206" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/2d102b65-bf8c-4df4-bca3-eadf749de206]: expected pod UID "2d102b65-bf8c-4df4-bca3-eadf749de206" but got "60736cbf-afd3-410e-87b8-bcf852ff2221" from Kube API Sep 29 10:41:41 crc kubenswrapper[4727]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Sep 29 10:41:41 crc kubenswrapper[4727]: > Sep 29 10:41:41 crc kubenswrapper[4727]: E0929 10:41:41.012728 4727 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Sep 29 10:41:41 crc kubenswrapper[4727]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_2d102b65-bf8c-4df4-bca3-eadf749de206_0(821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92" Netns:"/var/run/netns/a4127118-3f52-4ab7-9bfe-0b8e4d45486b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=821e0a0445154793311fc6f7eda4548fc13e0a2392d2b2c41ad1f3489580fa92;K8S_POD_UID=2d102b65-bf8c-4df4-bca3-eadf749de206" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/2d102b65-bf8c-4df4-bca3-eadf749de206]: expected pod UID "2d102b65-bf8c-4df4-bca3-eadf749de206" but got "60736cbf-afd3-410e-87b8-bcf852ff2221" from Kube API Sep 29 10:41:41 crc kubenswrapper[4727]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Sep 29 10:41:41 crc kubenswrapper[4727]: > pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.022998 4727 generic.go:334] "Generic (PLEG): container finished" podID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerID="03b780a0d9369fd3b6cd419c58c239b1ece3bf21c04aa3f5f396da48aa07503f" exitCode=143 Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.023083 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerDied","Data":"03b780a0d9369fd3b6cd419c58c239b1ece3bf21c04aa3f5f396da48aa07503f"} Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.024626 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gmmb\" (UniqueName: \"kubernetes.io/projected/60736cbf-afd3-410e-87b8-bcf852ff2221-kube-api-access-2gmmb\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.024668 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-combined-ca-bundle\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.024694 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.024742 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config-secret\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.028702 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.035981 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-openstack-config-secret\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.042165 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerStarted","Data":"d6fd27a8fb436f65cd0c919282146aa19379b0b7e1bd174e0e70b3486a6d8951"} Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.042212 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerStarted","Data":"02a8342c3dffaccd43c6894863031bac4a515f94bff86cb56efc687bd02cd084"} Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.049070 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60736cbf-afd3-410e-87b8-bcf852ff2221-combined-ca-bundle\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.068970 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gmmb\" (UniqueName: \"kubernetes.io/projected/60736cbf-afd3-410e-87b8-bcf852ff2221-kube-api-access-2gmmb\") pod \"openstackclient\" (UID: \"60736cbf-afd3-410e-87b8-bcf852ff2221\") " pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.119024 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.571203 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.641881 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle\") pod \"2ab43089-e386-4585-9324-79b4750eeeff\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.642238 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxzl8\" (UniqueName: \"kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8\") pod \"2ab43089-e386-4585-9324-79b4750eeeff\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.642257 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config\") pod \"2ab43089-e386-4585-9324-79b4750eeeff\" (UID: \"2ab43089-e386-4585-9324-79b4750eeeff\") " Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.658920 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8" (OuterVolumeSpecName: "kube-api-access-fxzl8") pod "2ab43089-e386-4585-9324-79b4750eeeff" (UID: "2ab43089-e386-4585-9324-79b4750eeeff"). InnerVolumeSpecName "kube-api-access-fxzl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.744009 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxzl8\" (UniqueName: \"kubernetes.io/projected/2ab43089-e386-4585-9324-79b4750eeeff-kube-api-access-fxzl8\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.757104 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ab43089-e386-4585-9324-79b4750eeeff" (UID: "2ab43089-e386-4585-9324-79b4750eeeff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.759413 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config" (OuterVolumeSpecName: "config") pod "2ab43089-e386-4585-9324-79b4750eeeff" (UID: "2ab43089-e386-4585-9324-79b4750eeeff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.777924 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.883217 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:41 crc kubenswrapper[4727]: I0929 10:41:41.883264 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ab43089-e386-4585-9324-79b4750eeeff-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.084399 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"60736cbf-afd3-410e-87b8-bcf852ff2221","Type":"ContainerStarted","Data":"61e7587fdf47a36e11cc4af9e28ffa1d588ac02e53f290ceb1e526a1d05f4236"} Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.090567 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8n57q" event={"ID":"2ab43089-e386-4585-9324-79b4750eeeff","Type":"ContainerDied","Data":"161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea"} Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.090719 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161716edbe7120d48c5c0dbfa28297ecd851f6924b6461cad5d430a0044203ea" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.090833 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8n57q" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.130635 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.131186 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerStarted","Data":"a8b8b42c467a683f9f2a738aac00cfe1e0537e9855ea6fbf44988006ee0fbb9e"} Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.178780 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.187064 4727 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d102b65-bf8c-4df4-bca3-eadf749de206" podUID="60736cbf-afd3-410e-87b8-bcf852ff2221" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.221431 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.221759 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="dnsmasq-dns" containerID="cri-o://63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c" gracePeriod=10 Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.242355 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:41:42 crc kubenswrapper[4727]: E0929 10:41:42.242993 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab43089-e386-4585-9324-79b4750eeeff" containerName="neutron-db-sync" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.243009 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab43089-e386-4585-9324-79b4750eeeff" containerName="neutron-db-sync" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.243195 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab43089-e386-4585-9324-79b4750eeeff" containerName="neutron-db-sync" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.244148 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.246816 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.246990 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-45t5g" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.248616 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.248754 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.264528 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.275542 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.279046 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.291703 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret\") pod \"2d102b65-bf8c-4df4-bca3-eadf749de206\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.291904 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle\") pod \"2d102b65-bf8c-4df4-bca3-eadf749de206\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.291946 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config\") pod \"2d102b65-bf8c-4df4-bca3-eadf749de206\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.296168 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nz6m\" (UniqueName: \"kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m\") pod \"2d102b65-bf8c-4df4-bca3-eadf749de206\" (UID: \"2d102b65-bf8c-4df4-bca3-eadf749de206\") " Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.296714 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhwcf\" (UniqueName: \"kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.296867 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.296926 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.297051 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.297100 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.304260 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d102b65-bf8c-4df4-bca3-eadf749de206" (UID: "2d102b65-bf8c-4df4-bca3-eadf749de206"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.305310 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.309029 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2d102b65-bf8c-4df4-bca3-eadf749de206" (UID: "2d102b65-bf8c-4df4-bca3-eadf749de206"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.314713 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2d102b65-bf8c-4df4-bca3-eadf749de206" (UID: "2d102b65-bf8c-4df4-bca3-eadf749de206"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.328505 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m" (OuterVolumeSpecName: "kube-api-access-8nz6m") pod "2d102b65-bf8c-4df4-bca3-eadf749de206" (UID: "2d102b65-bf8c-4df4-bca3-eadf749de206"). InnerVolumeSpecName "kube-api-access-8nz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.403721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhwcf\" (UniqueName: \"kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.403807 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404079 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404146 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404307 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404362 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404394 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404445 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404693 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404733 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404758 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ngzp\" (UniqueName: \"kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404832 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404850 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d102b65-bf8c-4df4-bca3-eadf749de206-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404862 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d102b65-bf8c-4df4-bca3-eadf749de206-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.404874 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nz6m\" (UniqueName: \"kubernetes.io/projected/2d102b65-bf8c-4df4-bca3-eadf749de206-kube-api-access-8nz6m\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.409733 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.410456 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.416022 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.422024 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.425998 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhwcf\" (UniqueName: \"kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf\") pod \"neutron-666c5cdbc4-zl8tq\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508675 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508805 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508835 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508925 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508955 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.508979 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ngzp\" (UniqueName: \"kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.509974 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.510730 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.510764 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.512144 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.513394 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.533062 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ngzp\" (UniqueName: \"kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp\") pod \"dnsmasq-dns-64667c4f57-m4xhc\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.645644 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.680316 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:42 crc kubenswrapper[4727]: I0929 10:41:42.928374 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.025914 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.026202 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.026282 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.026353 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.026467 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.026549 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcmnq\" (UniqueName: \"kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq\") pod \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\" (UID: \"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8\") " Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.034779 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq" (OuterVolumeSpecName: "kube-api-access-zcmnq") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "kube-api-access-zcmnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.135634 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcmnq\" (UniqueName: \"kubernetes.io/projected/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-kube-api-access-zcmnq\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.208732 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d102b65-bf8c-4df4-bca3-eadf749de206" path="/var/lib/kubelet/pods/2d102b65-bf8c-4df4-bca3-eadf749de206/volumes" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.258630 4727 generic.go:334] "Generic (PLEG): container finished" podID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerID="63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c" exitCode=0 Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.258729 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.281696 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.293498 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.337724 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config" (OuterVolumeSpecName: "config") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.341590 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.341619 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.352780 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.373421 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" event={"ID":"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8","Type":"ContainerDied","Data":"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c"} Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.373468 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85cb9df8d7-5dc48" event={"ID":"0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8","Type":"ContainerDied","Data":"c12faa2ead0ed9393b7fd110b003189f6acb3887b6f16a9e96c959cf2d9f9669"} Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.373479 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerStarted","Data":"be0f8915b4596fc40dfe20c98a39245372d74974ec49a3ae09015d8a3f8290e1"} Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.373499 4727 scope.go:117] "RemoveContainer" containerID="63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.382800 4727 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d102b65-bf8c-4df4-bca3-eadf749de206" podUID="60736cbf-afd3-410e-87b8-bcf852ff2221" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.384993 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.436713 4727 scope.go:117] "RemoveContainer" containerID="0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.437450 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" (UID: "0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.444562 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.444595 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.444607 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.511390 4727 scope.go:117] "RemoveContainer" containerID="63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c" Sep 29 10:41:43 crc kubenswrapper[4727]: E0929 10:41:43.520521 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c\": container with ID starting with 63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c not found: ID does not exist" containerID="63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.520573 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c"} err="failed to get container status \"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c\": rpc error: code = NotFound desc = could not find container \"63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c\": container with ID starting with 63aad5e65dabbb65b24b2f0ae2476b4c0e04caa417c058c0ef3c3580a0cc7a2c not found: ID does not exist" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.520606 4727 scope.go:117] "RemoveContainer" containerID="0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993" Sep 29 10:41:43 crc kubenswrapper[4727]: E0929 10:41:43.524482 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993\": container with ID starting with 0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993 not found: ID does not exist" containerID="0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.524536 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993"} err="failed to get container status \"0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993\": rpc error: code = NotFound desc = could not find container \"0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993\": container with ID starting with 0cf718e420961fb2e36d5c6342269b076366bcc8e6f346544817fbca0ae7b993 not found: ID does not exist" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.528206 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.553676 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.638623 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.650558 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85cb9df8d7-5dc48"] Sep 29 10:41:43 crc kubenswrapper[4727]: I0929 10:41:43.943203 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:41:43 crc kubenswrapper[4727]: W0929 10:41:43.955425 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f6bd456_e04d_4dcc_9c57_d3f8e7c9919a.slice/crio-506e7dbef15528cca921ab2e9016a81743c2931693174f0ae3b56cdb05835c27 WatchSource:0}: Error finding container 506e7dbef15528cca921ab2e9016a81743c2931693174f0ae3b56cdb05835c27: Status 404 returned error can't find the container with id 506e7dbef15528cca921ab2e9016a81743c2931693174f0ae3b56cdb05835c27 Sep 29 10:41:44 crc kubenswrapper[4727]: I0929 10:41:44.295033 4727 generic.go:334] "Generic (PLEG): container finished" podID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerID="ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37" exitCode=0 Sep 29 10:41:44 crc kubenswrapper[4727]: I0929 10:41:44.295093 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" event={"ID":"17498e09-aa36-4ba6-8942-f8c18bdc8f78","Type":"ContainerDied","Data":"ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37"} Sep 29 10:41:44 crc kubenswrapper[4727]: I0929 10:41:44.295209 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" event={"ID":"17498e09-aa36-4ba6-8942-f8c18bdc8f78","Type":"ContainerStarted","Data":"690ed332b4a63dd6b7e150e456e15e288b04fa2e5861efe46e6b2e4ae6194fac"} Sep 29 10:41:44 crc kubenswrapper[4727]: I0929 10:41:44.297668 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerStarted","Data":"4845921ebcc5c86479aa47e1c246cdc0284636f6e5280acad46fd47666235adf"} Sep 29 10:41:44 crc kubenswrapper[4727]: I0929 10:41:44.297720 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerStarted","Data":"506e7dbef15528cca921ab2e9016a81743c2931693174f0ae3b56cdb05835c27"} Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.129675 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" path="/var/lib/kubelet/pods/0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8/volumes" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.312330 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerStarted","Data":"ae72109ce362862b0c543a6a98284e2ba0c67c420ef27519665a0e4e0fd27d94"} Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.312977 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.317072 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerStarted","Data":"718a7ccd5145c6b46e5c9714ed7a7340f6ccdf101fb99acfc80bec8dd46b4371"} Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.317268 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.321637 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" event={"ID":"17498e09-aa36-4ba6-8942-f8c18bdc8f78","Type":"ContainerStarted","Data":"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484"} Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.321809 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.349320 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-666c5cdbc4-zl8tq" podStartSLOduration=3.349297084 podStartE2EDuration="3.349297084s" podCreationTimestamp="2025-09-29 10:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:45.330199393 +0000 UTC m=+1175.503512755" watchObservedRunningTime="2025-09-29 10:41:45.349297084 +0000 UTC m=+1175.522610446" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.359269 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.967351683 podStartE2EDuration="7.359252721s" podCreationTimestamp="2025-09-29 10:41:38 +0000 UTC" firstStartedPulling="2025-09-29 10:41:39.994701407 +0000 UTC m=+1170.168014769" lastFinishedPulling="2025-09-29 10:41:44.386602445 +0000 UTC m=+1174.559915807" observedRunningTime="2025-09-29 10:41:45.354042226 +0000 UTC m=+1175.527355588" watchObservedRunningTime="2025-09-29 10:41:45.359252721 +0000 UTC m=+1175.532566083" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.378788 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" podStartSLOduration=3.378768833 podStartE2EDuration="3.378768833s" podCreationTimestamp="2025-09-29 10:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:45.374821191 +0000 UTC m=+1175.548134553" watchObservedRunningTime="2025-09-29 10:41:45.378768833 +0000 UTC m=+1175.552082195" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.635531 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.689167 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68ff49cc97-zxm88"] Sep 29 10:41:45 crc kubenswrapper[4727]: E0929 10:41:45.689575 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="dnsmasq-dns" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.689590 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="dnsmasq-dns" Sep 29 10:41:45 crc kubenswrapper[4727]: E0929 10:41:45.689615 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="init" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.689622 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="init" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.689798 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac2ffa4-3d80-4119-abb3-0a91ed4e3ba8" containerName="dnsmasq-dns" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.690733 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.693376 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.693867 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.699690 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68ff49cc97-zxm88"] Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821636 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-public-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821683 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-ovndb-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821744 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-internal-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821767 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twxql\" (UniqueName: \"kubernetes.io/projected/326c2e7e-51f1-4f8c-9d91-4572393d8c11-kube-api-access-twxql\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821809 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821848 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-httpd-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.821947 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-combined-ca-bundle\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923619 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-internal-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923663 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twxql\" (UniqueName: \"kubernetes.io/projected/326c2e7e-51f1-4f8c-9d91-4572393d8c11-kube-api-access-twxql\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923721 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923762 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-httpd-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923780 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-combined-ca-bundle\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923834 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-public-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.923859 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-ovndb-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.928946 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-internal-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.930638 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-ovndb-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.934025 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-combined-ca-bundle\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.934255 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.934792 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-public-tls-certs\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.936439 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/326c2e7e-51f1-4f8c-9d91-4572393d8c11-httpd-config\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:45 crc kubenswrapper[4727]: I0929 10:41:45.947866 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twxql\" (UniqueName: \"kubernetes.io/projected/326c2e7e-51f1-4f8c-9d91-4572393d8c11-kube-api-access-twxql\") pod \"neutron-68ff49cc97-zxm88\" (UID: \"326c2e7e-51f1-4f8c-9d91-4572393d8c11\") " pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:46 crc kubenswrapper[4727]: I0929 10:41:46.013387 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:46 crc kubenswrapper[4727]: I0929 10:41:46.588260 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68ff49cc97-zxm88"] Sep 29 10:41:46 crc kubenswrapper[4727]: W0929 10:41:46.607777 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod326c2e7e_51f1_4f8c_9d91_4572393d8c11.slice/crio-1b153a5a3c0f24203b152aab35e6844425c26b715c702d43227349483cdde658 WatchSource:0}: Error finding container 1b153a5a3c0f24203b152aab35e6844425c26b715c702d43227349483cdde658: Status 404 returned error can't find the container with id 1b153a5a3c0f24203b152aab35e6844425c26b715c702d43227349483cdde658 Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.351447 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ff49cc97-zxm88" event={"ID":"326c2e7e-51f1-4f8c-9d91-4572393d8c11","Type":"ContainerStarted","Data":"9a828a292cc7617c8df47cdbeac5e55b5b27dbea72ffe26a2b985446a4a21a8b"} Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.351859 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ff49cc97-zxm88" event={"ID":"326c2e7e-51f1-4f8c-9d91-4572393d8c11","Type":"ContainerStarted","Data":"bd8598489b2851667b16ee68d5f0cc40c25ab6e53740d70540e2ceb581ca7b3c"} Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.351873 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ff49cc97-zxm88" event={"ID":"326c2e7e-51f1-4f8c-9d91-4572393d8c11","Type":"ContainerStarted","Data":"1b153a5a3c0f24203b152aab35e6844425c26b715c702d43227349483cdde658"} Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.351892 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.356388 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sg2mp" event={"ID":"62cb5972-95c1-4963-b0f2-bfee701d6453","Type":"ContainerStarted","Data":"0a94d38d0c451c27ed76e6ae16ddd0fd73b4aa3755c57a5fc2e041bfa37227fa"} Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.378078 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68ff49cc97-zxm88" podStartSLOduration=2.378060671 podStartE2EDuration="2.378060671s" podCreationTimestamp="2025-09-29 10:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:47.372521469 +0000 UTC m=+1177.545834841" watchObservedRunningTime="2025-09-29 10:41:47.378060671 +0000 UTC m=+1177.551374033" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.401239 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-sg2mp" podStartSLOduration=8.44605945 podStartE2EDuration="44.401213187s" podCreationTimestamp="2025-09-29 10:41:03 +0000 UTC" firstStartedPulling="2025-09-29 10:41:09.816817002 +0000 UTC m=+1139.990130364" lastFinishedPulling="2025-09-29 10:41:45.771970739 +0000 UTC m=+1175.945284101" observedRunningTime="2025-09-29 10:41:47.393519469 +0000 UTC m=+1177.566832841" watchObservedRunningTime="2025-09-29 10:41:47.401213187 +0000 UTC m=+1177.574526549" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.786722 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-598db7f995-m4hgm"] Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.789228 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.796696 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.796960 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.797075 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.815712 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-598db7f995-m4hgm"] Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.864499 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-etc-swift\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.864894 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-config-data\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.864913 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-internal-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.864934 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-run-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.864986 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmnh4\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-kube-api-access-gmnh4\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.865028 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-public-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.865052 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-log-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.865069 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-combined-ca-bundle\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966491 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmnh4\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-kube-api-access-gmnh4\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966580 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-public-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966621 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-log-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966647 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-combined-ca-bundle\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966734 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-etc-swift\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966782 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-config-data\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966804 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-internal-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.966827 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-run-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.967406 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-log-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.967442 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-run-httpd\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.980030 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-config-data\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.985032 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-internal-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.987288 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-etc-swift\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.987702 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-combined-ca-bundle\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.988282 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmnh4\" (UniqueName: \"kubernetes.io/projected/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-kube-api-access-gmnh4\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:47 crc kubenswrapper[4727]: I0929 10:41:47.989952 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87ecfca9-38cc-4b74-88b7-2d56d8f5638d-public-tls-certs\") pod \"swift-proxy-598db7f995-m4hgm\" (UID: \"87ecfca9-38cc-4b74-88b7-2d56d8f5638d\") " pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.127569 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.140922 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:34584->10.217.0.159:9311: read: connection reset by peer" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.140933 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b84775bb-fqb4p" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:34588->10.217.0.159:9311: read: connection reset by peer" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.381061 4727 generic.go:334] "Generic (PLEG): container finished" podID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerID="0bfdacb4ed15db81f9496ff96bd31129b7c79e1e82a92bc6231ee0a95c6b0e68" exitCode=0 Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.381916 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerDied","Data":"0bfdacb4ed15db81f9496ff96bd31129b7c79e1e82a92bc6231ee0a95c6b0e68"} Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.776654 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.782176 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-598db7f995-m4hgm"] Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.843621 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.844186 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-central-agent" containerID="cri-o://d6fd27a8fb436f65cd0c919282146aa19379b0b7e1bd174e0e70b3486a6d8951" gracePeriod=30 Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.844647 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-notification-agent" containerID="cri-o://a8b8b42c467a683f9f2a738aac00cfe1e0537e9855ea6fbf44988006ee0fbb9e" gracePeriod=30 Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.844720 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="sg-core" containerID="cri-o://be0f8915b4596fc40dfe20c98a39245372d74974ec49a3ae09015d8a3f8290e1" gracePeriod=30 Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.844915 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="proxy-httpd" containerID="cri-o://718a7ccd5145c6b46e5c9714ed7a7340f6ccdf101fb99acfc80bec8dd46b4371" gracePeriod=30 Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.885380 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data\") pod \"bd673ffa-5d52-490b-8065-36767a0b53cf\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.885505 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs\") pod \"bd673ffa-5d52-490b-8065-36767a0b53cf\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.885550 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom\") pod \"bd673ffa-5d52-490b-8065-36767a0b53cf\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.885624 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle\") pod \"bd673ffa-5d52-490b-8065-36767a0b53cf\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.885718 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tspq4\" (UniqueName: \"kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4\") pod \"bd673ffa-5d52-490b-8065-36767a0b53cf\" (UID: \"bd673ffa-5d52-490b-8065-36767a0b53cf\") " Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.886321 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs" (OuterVolumeSpecName: "logs") pod "bd673ffa-5d52-490b-8065-36767a0b53cf" (UID: "bd673ffa-5d52-490b-8065-36767a0b53cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.890987 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bd673ffa-5d52-490b-8065-36767a0b53cf" (UID: "bd673ffa-5d52-490b-8065-36767a0b53cf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.892286 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4" (OuterVolumeSpecName: "kube-api-access-tspq4") pod "bd673ffa-5d52-490b-8065-36767a0b53cf" (UID: "bd673ffa-5d52-490b-8065-36767a0b53cf"). InnerVolumeSpecName "kube-api-access-tspq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.917251 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd673ffa-5d52-490b-8065-36767a0b53cf" (UID: "bd673ffa-5d52-490b-8065-36767a0b53cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.970173 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data" (OuterVolumeSpecName: "config-data") pod "bd673ffa-5d52-490b-8065-36767a0b53cf" (UID: "bd673ffa-5d52-490b-8065-36767a0b53cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.987797 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd673ffa-5d52-490b-8065-36767a0b53cf-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.987856 4727 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.987869 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.987880 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tspq4\" (UniqueName: \"kubernetes.io/projected/bd673ffa-5d52-490b-8065-36767a0b53cf-kube-api-access-tspq4\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:48 crc kubenswrapper[4727]: I0929 10:41:48.987892 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd673ffa-5d52-490b-8065-36767a0b53cf-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.268131 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.268466 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.268541 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.269321 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.269493 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b" gracePeriod=600 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.407101 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b" exitCode=0 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.407165 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.407198 4727 scope.go:117] "RemoveContainer" containerID="f103339e5cd6894b8747a94eb003e2ea7de14aefdf85677a5f686a4bb2b22435" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.412607 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b84775bb-fqb4p" event={"ID":"bd673ffa-5d52-490b-8065-36767a0b53cf","Type":"ContainerDied","Data":"3182d4cbfa73fc7ce7c154132f3bc54cdebde1142de39b85534186d9eaf19553"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.412689 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b84775bb-fqb4p" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.417671 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-598db7f995-m4hgm" event={"ID":"87ecfca9-38cc-4b74-88b7-2d56d8f5638d","Type":"ContainerStarted","Data":"adaac5e7fe0b170f946635f47c80637e6bd779e936f3ffc776ae85b2e39a49ec"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.417718 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-598db7f995-m4hgm" event={"ID":"87ecfca9-38cc-4b74-88b7-2d56d8f5638d","Type":"ContainerStarted","Data":"0dc4e5cea53bf5dc7a7805f24bc623fc7c708b305da796ce7e55aade604c3e0a"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.417735 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-598db7f995-m4hgm" event={"ID":"87ecfca9-38cc-4b74-88b7-2d56d8f5638d","Type":"ContainerStarted","Data":"112373a5d74810aca46f60b7bf812e0fcfcc3936fef955a19c47c989ef3eaa7e"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.417942 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.417963 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436047 4727 generic.go:334] "Generic (PLEG): container finished" podID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerID="718a7ccd5145c6b46e5c9714ed7a7340f6ccdf101fb99acfc80bec8dd46b4371" exitCode=0 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436080 4727 generic.go:334] "Generic (PLEG): container finished" podID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerID="be0f8915b4596fc40dfe20c98a39245372d74974ec49a3ae09015d8a3f8290e1" exitCode=2 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436088 4727 generic.go:334] "Generic (PLEG): container finished" podID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerID="a8b8b42c467a683f9f2a738aac00cfe1e0537e9855ea6fbf44988006ee0fbb9e" exitCode=0 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436096 4727 generic.go:334] "Generic (PLEG): container finished" podID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerID="d6fd27a8fb436f65cd0c919282146aa19379b0b7e1bd174e0e70b3486a6d8951" exitCode=0 Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436119 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerDied","Data":"718a7ccd5145c6b46e5c9714ed7a7340f6ccdf101fb99acfc80bec8dd46b4371"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436144 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerDied","Data":"be0f8915b4596fc40dfe20c98a39245372d74974ec49a3ae09015d8a3f8290e1"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436154 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerDied","Data":"a8b8b42c467a683f9f2a738aac00cfe1e0537e9855ea6fbf44988006ee0fbb9e"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.436162 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerDied","Data":"d6fd27a8fb436f65cd0c919282146aa19379b0b7e1bd174e0e70b3486a6d8951"} Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.443440 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-598db7f995-m4hgm" podStartSLOduration=2.443422171 podStartE2EDuration="2.443422171s" podCreationTimestamp="2025-09-29 10:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:49.438067583 +0000 UTC m=+1179.611380945" watchObservedRunningTime="2025-09-29 10:41:49.443422171 +0000 UTC m=+1179.616735533" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.468104 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.474480 4727 scope.go:117] "RemoveContainer" containerID="0bfdacb4ed15db81f9496ff96bd31129b7c79e1e82a92bc6231ee0a95c6b0e68" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.476228 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-64b84775bb-fqb4p"] Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.515789 4727 scope.go:117] "RemoveContainer" containerID="03b780a0d9369fd3b6cd419c58c239b1ece3bf21c04aa3f5f396da48aa07503f" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.800109 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.858970 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.910678 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911466 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911475 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911518 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911649 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911680 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911714 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn24b\" (UniqueName: \"kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.911765 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml\") pod \"4f0f2499-9305-48e2-828b-28075cd9e7ee\" (UID: \"4f0f2499-9305-48e2-828b-28075cd9e7ee\") " Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.912272 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.912612 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.920189 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b" (OuterVolumeSpecName: "kube-api-access-tn24b") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "kube-api-access-tn24b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.926501 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts" (OuterVolumeSpecName: "scripts") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:49 crc kubenswrapper[4727]: I0929 10:41:49.956005 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.012603 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.013997 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.014022 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.014032 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f0f2499-9305-48e2-828b-28075cd9e7ee-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.014040 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn24b\" (UniqueName: \"kubernetes.io/projected/4f0f2499-9305-48e2-828b-28075cd9e7ee-kube-api-access-tn24b\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.014050 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.024616 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data" (OuterVolumeSpecName: "config-data") pod "4f0f2499-9305-48e2-828b-28075cd9e7ee" (UID: "4f0f2499-9305-48e2-828b-28075cd9e7ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.120460 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0f2499-9305-48e2-828b-28075cd9e7ee-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.450783 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f0f2499-9305-48e2-828b-28075cd9e7ee","Type":"ContainerDied","Data":"02a8342c3dffaccd43c6894863031bac4a515f94bff86cb56efc687bd02cd084"} Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.451113 4727 scope.go:117] "RemoveContainer" containerID="718a7ccd5145c6b46e5c9714ed7a7340f6ccdf101fb99acfc80bec8dd46b4371" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.451069 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.456326 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315"} Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.513403 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.522857 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531093 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531595 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-central-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531610 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-central-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531629 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531639 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531647 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531653 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531670 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-notification-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531675 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-notification-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531697 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="proxy-httpd" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531703 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="proxy-httpd" Sep 29 10:41:50 crc kubenswrapper[4727]: E0929 10:41:50.531722 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="sg-core" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531729 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="sg-core" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531938 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="proxy-httpd" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531958 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531970 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-central-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531979 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="sg-core" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.531991 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" containerName="barbican-api-log" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.532002 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" containerName="ceilometer-notification-agent" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.533772 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.537936 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.538156 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.541534 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631612 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631679 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631722 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631751 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631770 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631798 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjzs2\" (UniqueName: \"kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.631838 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733255 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733355 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733401 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733425 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733462 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjzs2\" (UniqueName: \"kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733517 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.733567 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.736110 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.736384 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.743720 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.746117 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.748836 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.761071 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.771067 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjzs2\" (UniqueName: \"kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2\") pod \"ceilometer-0\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " pod="openstack/ceilometer-0" Sep 29 10:41:50 crc kubenswrapper[4727]: I0929 10:41:50.874064 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:41:51 crc kubenswrapper[4727]: I0929 10:41:51.120898 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f0f2499-9305-48e2-828b-28075cd9e7ee" path="/var/lib/kubelet/pods/4f0f2499-9305-48e2-828b-28075cd9e7ee/volumes" Sep 29 10:41:51 crc kubenswrapper[4727]: I0929 10:41:51.122626 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd673ffa-5d52-490b-8065-36767a0b53cf" path="/var/lib/kubelet/pods/bd673ffa-5d52-490b-8065-36767a0b53cf/volumes" Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.683516 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.769707 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.769961 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="dnsmasq-dns" containerID="cri-o://066b43bb1594b6e4a204ccd6714333c0c1e6cd6ba474b75b7da7e62b416f6c26" gracePeriod=10 Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.832792 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-l6trr"] Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.841158 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l6trr" Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.848947 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-l6trr"] Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.916507 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-rrbt8"] Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.919662 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.943606 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rrbt8"] Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.981880 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk7q5\" (UniqueName: \"kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5\") pod \"nova-cell0-db-create-rrbt8\" (UID: \"77b2d573-abbb-491f-8939-6c7f79082906\") " pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:41:52 crc kubenswrapper[4727]: I0929 10:41:52.982060 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kddr\" (UniqueName: \"kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr\") pod \"nova-api-db-create-l6trr\" (UID: \"6be32091-cf3a-4931-bcb8-2c867604774e\") " pod="openstack/nova-api-db-create-l6trr" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.014525 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dbvq2"] Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.015760 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.034792 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dbvq2"] Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.085445 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kddr\" (UniqueName: \"kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr\") pod \"nova-api-db-create-l6trr\" (UID: \"6be32091-cf3a-4931-bcb8-2c867604774e\") " pod="openstack/nova-api-db-create-l6trr" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.085531 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kpq7\" (UniqueName: \"kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7\") pod \"nova-cell1-db-create-dbvq2\" (UID: \"973c7613-3c61-4534-b26e-e1d0ea649bc9\") " pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.085561 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk7q5\" (UniqueName: \"kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5\") pod \"nova-cell0-db-create-rrbt8\" (UID: \"77b2d573-abbb-491f-8939-6c7f79082906\") " pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.106428 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kddr\" (UniqueName: \"kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr\") pod \"nova-api-db-create-l6trr\" (UID: \"6be32091-cf3a-4931-bcb8-2c867604774e\") " pod="openstack/nova-api-db-create-l6trr" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.107278 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk7q5\" (UniqueName: \"kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5\") pod \"nova-cell0-db-create-rrbt8\" (UID: \"77b2d573-abbb-491f-8939-6c7f79082906\") " pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.174135 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l6trr" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.189324 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kpq7\" (UniqueName: \"kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7\") pod \"nova-cell1-db-create-dbvq2\" (UID: \"973c7613-3c61-4534-b26e-e1d0ea649bc9\") " pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.212772 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kpq7\" (UniqueName: \"kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7\") pod \"nova-cell1-db-create-dbvq2\" (UID: \"973c7613-3c61-4534-b26e-e1d0ea649bc9\") " pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.259971 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.339118 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.485300 4727 generic.go:334] "Generic (PLEG): container finished" podID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerID="066b43bb1594b6e4a204ccd6714333c0c1e6cd6ba474b75b7da7e62b416f6c26" exitCode=0 Sep 29 10:41:53 crc kubenswrapper[4727]: I0929 10:41:53.485379 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" event={"ID":"82a82880-cb7a-40e1-9908-03f142f6ff71","Type":"ContainerDied","Data":"066b43bb1594b6e4a204ccd6714333c0c1e6cd6ba474b75b7da7e62b416f6c26"} Sep 29 10:41:54 crc kubenswrapper[4727]: I0929 10:41:54.496207 4727 generic.go:334] "Generic (PLEG): container finished" podID="62cb5972-95c1-4963-b0f2-bfee701d6453" containerID="0a94d38d0c451c27ed76e6ae16ddd0fd73b4aa3755c57a5fc2e041bfa37227fa" exitCode=0 Sep 29 10:41:54 crc kubenswrapper[4727]: I0929 10:41:54.496462 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sg2mp" event={"ID":"62cb5972-95c1-4963-b0f2-bfee701d6453","Type":"ContainerDied","Data":"0a94d38d0c451c27ed76e6ae16ddd0fd73b4aa3755c57a5fc2e041bfa37227fa"} Sep 29 10:41:54 crc kubenswrapper[4727]: I0929 10:41:54.727152 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.389744 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.427162 4727 scope.go:117] "RemoveContainer" containerID="be0f8915b4596fc40dfe20c98a39245372d74974ec49a3ae09015d8a3f8290e1" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455059 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455149 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455208 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455311 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgwjr\" (UniqueName: \"kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455406 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.455438 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data\") pod \"62cb5972-95c1-4963-b0f2-bfee701d6453\" (UID: \"62cb5972-95c1-4963-b0f2-bfee701d6453\") " Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.456536 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.470420 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.470761 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr" (OuterVolumeSpecName: "kube-api-access-zgwjr") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "kube-api-access-zgwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.480030 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts" (OuterVolumeSpecName: "scripts") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.514850 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.546501 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data" (OuterVolumeSpecName: "config-data") pod "62cb5972-95c1-4963-b0f2-bfee701d6453" (UID: "62cb5972-95c1-4963-b0f2-bfee701d6453"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.551613 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sg2mp" event={"ID":"62cb5972-95c1-4963-b0f2-bfee701d6453","Type":"ContainerDied","Data":"0e37e5709f7ca6384260459d6dfd3c5cff249d90a8d796ac047c19791082e8b2"} Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.551649 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e37e5709f7ca6384260459d6dfd3c5cff249d90a8d796ac047c19791082e8b2" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.551705 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sg2mp" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557205 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgwjr\" (UniqueName: \"kubernetes.io/projected/62cb5972-95c1-4963-b0f2-bfee701d6453-kube-api-access-zgwjr\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557236 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557246 4727 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557257 4727 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62cb5972-95c1-4963-b0f2-bfee701d6453-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557264 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.557272 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb5972-95c1-4963-b0f2-bfee701d6453-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.733149 4727 scope.go:117] "RemoveContainer" containerID="a8b8b42c467a683f9f2a738aac00cfe1e0537e9855ea6fbf44988006ee0fbb9e" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.767718 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:41:56 crc kubenswrapper[4727]: E0929 10:41:56.768224 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" containerName="cinder-db-sync" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.768241 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" containerName="cinder-db-sync" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.768545 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" containerName="cinder-db-sync" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.769818 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.791292 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.810690 4727 scope.go:117] "RemoveContainer" containerID="d6fd27a8fb436f65cd0c919282146aa19379b0b7e1bd174e0e70b3486a6d8951" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.872013 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.883954 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhm5m\" (UniqueName: \"kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.884093 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.884195 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.884239 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.884363 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.901854 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.924874 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.936558 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="dnsmasq-dns" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.937864 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.938542 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.940808 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7bgkq" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.941736 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.952331 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.952542 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.986823 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhm5m\" (UniqueName: \"kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.986893 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.986939 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.986961 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.987009 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.987075 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:56 crc kubenswrapper[4727]: I0929 10:41:56.987826 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:56.997612 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:56.997657 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:56.997733 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.020400 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhm5m\" (UniqueName: \"kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.028818 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0\") pod \"dnsmasq-dns-65bc8f75b9-xf5sx\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.061408 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:41:57 crc kubenswrapper[4727]: E0929 10:41:57.061785 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="dnsmasq-dns" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.061800 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="dnsmasq-dns" Sep 29 10:41:57 crc kubenswrapper[4727]: E0929 10:41:57.061815 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="init" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.061821 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" containerName="init" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.062960 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.065759 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.081450 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.090361 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsf98\" (UniqueName: \"kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98\") pod \"82a82880-cb7a-40e1-9908-03f142f6ff71\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.090634 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config\") pod \"82a82880-cb7a-40e1-9908-03f142f6ff71\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.090729 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc\") pod \"82a82880-cb7a-40e1-9908-03f142f6ff71\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.090982 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb\") pod \"82a82880-cb7a-40e1-9908-03f142f6ff71\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.091409 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb\") pod \"82a82880-cb7a-40e1-9908-03f142f6ff71\" (UID: \"82a82880-cb7a-40e1-9908-03f142f6ff71\") " Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.091755 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.092140 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.092217 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.092351 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.092645 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9866\" (UniqueName: \"kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.092729 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.097268 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98" (OuterVolumeSpecName: "kube-api-access-gsf98") pod "82a82880-cb7a-40e1-9908-03f142f6ff71" (UID: "82a82880-cb7a-40e1-9908-03f142f6ff71"). InnerVolumeSpecName "kube-api-access-gsf98". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.100137 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.193214 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194772 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194802 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nz2d\" (UniqueName: \"kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194835 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194856 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194872 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.194894 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195211 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195324 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195385 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195413 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195493 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195537 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9866\" (UniqueName: \"kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.195560 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.200834 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsf98\" (UniqueName: \"kubernetes.io/projected/82a82880-cb7a-40e1-9908-03f142f6ff71-kube-api-access-gsf98\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.201323 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.204636 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82a82880-cb7a-40e1-9908-03f142f6ff71" (UID: "82a82880-cb7a-40e1-9908-03f142f6ff71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.208972 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.212530 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.215459 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.219736 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.229760 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9866\" (UniqueName: \"kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866\") pod \"cinder-scheduler-0\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.266786 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config" (OuterVolumeSpecName: "config") pod "82a82880-cb7a-40e1-9908-03f142f6ff71" (UID: "82a82880-cb7a-40e1-9908-03f142f6ff71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.298583 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82a82880-cb7a-40e1-9908-03f142f6ff71" (UID: "82a82880-cb7a-40e1-9908-03f142f6ff71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302583 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302627 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nz2d\" (UniqueName: \"kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302663 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302680 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302839 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302874 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.302897 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.303049 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.303061 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.303070 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.304300 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.305738 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.313607 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.314946 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.332052 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.348972 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nz2d\" (UniqueName: \"kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.350113 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.351284 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data\") pod \"cinder-api-0\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.391482 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82a82880-cb7a-40e1-9908-03f142f6ff71" (UID: "82a82880-cb7a-40e1-9908-03f142f6ff71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.420953 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a82880-cb7a-40e1-9908-03f142f6ff71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.427746 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.526420 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-l6trr"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.538584 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rrbt8"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.588233 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerStarted","Data":"426c653ea90f98db398dd2c1691b49dfc74b9ecb04f9609ea64271a6d8f79ea4"} Sep 29 10:41:57 crc kubenswrapper[4727]: W0929 10:41:57.590545 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77b2d573_abbb_491f_8939_6c7f79082906.slice/crio-bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754 WatchSource:0}: Error finding container bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754: Status 404 returned error can't find the container with id bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754 Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.592260 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l6trr" event={"ID":"6be32091-cf3a-4931-bcb8-2c867604774e","Type":"ContainerStarted","Data":"4db178844fe8130e66b5f25d7cf33b02abfd2f0fec945fbc81de6fd7a35509f3"} Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.602477 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" event={"ID":"82a82880-cb7a-40e1-9908-03f142f6ff71","Type":"ContainerDied","Data":"53a1d0c9fca2d98abe2fd625c68cde7901dbe57fda1f164ebda3ad4c4594d139"} Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.602536 4727 scope.go:117] "RemoveContainer" containerID="066b43bb1594b6e4a204ccd6714333c0c1e6cd6ba474b75b7da7e62b416f6c26" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.602706 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-sd6r4" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.717284 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dbvq2"] Sep 29 10:41:57 crc kubenswrapper[4727]: W0929 10:41:57.744905 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod973c7613_3c61_4534_b26e_e1d0ea649bc9.slice/crio-4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a WatchSource:0}: Error finding container 4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a: Status 404 returned error can't find the container with id 4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.775508 4727 scope.go:117] "RemoveContainer" containerID="f6d8695459b715a3428091ccc94dc8da5976ee2ee61dcbfdf77916c3a3e7a864" Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.801142 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.809180 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-sd6r4"] Sep 29 10:41:57 crc kubenswrapper[4727]: I0929 10:41:57.898709 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.056183 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.117166 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.141148 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.145058 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-598db7f995-m4hgm" Sep 29 10:41:58 crc kubenswrapper[4727]: W0929 10:41:58.152466 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf74dfd8_697e_477a_82fb_62b76c4a18a4.slice/crio-2ab5a70892d899c39539ea4622e709243326a077029500f69450f8ddb9ac16eb WatchSource:0}: Error finding container 2ab5a70892d899c39539ea4622e709243326a077029500f69450f8ddb9ac16eb: Status 404 returned error can't find the container with id 2ab5a70892d899c39539ea4622e709243326a077029500f69450f8ddb9ac16eb Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.660177 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerStarted","Data":"2ab5a70892d899c39539ea4622e709243326a077029500f69450f8ddb9ac16eb"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.666371 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerStarted","Data":"962101826d72f2be1e526162a08e12be220d164a1232c2acd9d71e206d770833"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.669706 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rrbt8" event={"ID":"77b2d573-abbb-491f-8939-6c7f79082906","Type":"ContainerStarted","Data":"ff3a5688d5e995bc23e2c566cee352cfc332a17885d01f72318a20848b544e2f"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.669803 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rrbt8" event={"ID":"77b2d573-abbb-491f-8939-6c7f79082906","Type":"ContainerStarted","Data":"bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.673361 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l6trr" event={"ID":"6be32091-cf3a-4931-bcb8-2c867604774e","Type":"ContainerDied","Data":"ed760ac0fa0b366f35c1ecce6501e8be5a490d5ac39a5aac4e37816fdd40d529"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.673307 4727 generic.go:334] "Generic (PLEG): container finished" podID="6be32091-cf3a-4931-bcb8-2c867604774e" containerID="ed760ac0fa0b366f35c1ecce6501e8be5a490d5ac39a5aac4e37816fdd40d529" exitCode=0 Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.676096 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"60736cbf-afd3-410e-87b8-bcf852ff2221","Type":"ContainerStarted","Data":"20c619f96ed778679f93a52087ca5f19775e9d49512c6531441c6f1c0f5a283d"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.688958 4727 generic.go:334] "Generic (PLEG): container finished" podID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerID="6f81c50977d0a38667578e1d0d80b7454bf7cad0b4ee7c87621786c0ab0f6680" exitCode=0 Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.689077 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" event={"ID":"8150baa4-c7ed-4f99-9c0b-988c41630c97","Type":"ContainerDied","Data":"6f81c50977d0a38667578e1d0d80b7454bf7cad0b4ee7c87621786c0ab0f6680"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.689105 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" event={"ID":"8150baa4-c7ed-4f99-9c0b-988c41630c97","Type":"ContainerStarted","Data":"14c29fd90f304ca9f3ba72ffd53e4f7a02107b1bfff7aefb5b5f5338690bac8c"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.706796 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.741162418 podStartE2EDuration="18.706780744s" podCreationTimestamp="2025-09-29 10:41:40 +0000 UTC" firstStartedPulling="2025-09-29 10:41:41.850627838 +0000 UTC m=+1172.023941200" lastFinishedPulling="2025-09-29 10:41:56.816246174 +0000 UTC m=+1186.989559526" observedRunningTime="2025-09-29 10:41:58.702933985 +0000 UTC m=+1188.876247347" watchObservedRunningTime="2025-09-29 10:41:58.706780744 +0000 UTC m=+1188.880094106" Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.706947 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbvq2" event={"ID":"973c7613-3c61-4534-b26e-e1d0ea649bc9","Type":"ContainerStarted","Data":"24f9c8d5a43eabb8016a1b0fa2a9548dfe15ff4f232d0bb677913c0c30063dac"} Sep 29 10:41:58 crc kubenswrapper[4727]: I0929 10:41:58.706981 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbvq2" event={"ID":"973c7613-3c61-4534-b26e-e1d0ea649bc9","Type":"ContainerStarted","Data":"4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.118971 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a82880-cb7a-40e1-9908-03f142f6ff71" path="/var/lib/kubelet/pods/82a82880-cb7a-40e1-9908-03f142f6ff71/volumes" Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.580384 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.740711 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" event={"ID":"8150baa4-c7ed-4f99-9c0b-988c41630c97","Type":"ContainerStarted","Data":"c815a6292fd973c9eb2dc88428dbf3fcf81f7eaa88dc5f9db402c1ae9b7f899b"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.741575 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.742986 4727 generic.go:334] "Generic (PLEG): container finished" podID="973c7613-3c61-4534-b26e-e1d0ea649bc9" containerID="24f9c8d5a43eabb8016a1b0fa2a9548dfe15ff4f232d0bb677913c0c30063dac" exitCode=0 Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.743174 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbvq2" event={"ID":"973c7613-3c61-4534-b26e-e1d0ea649bc9","Type":"ContainerDied","Data":"24f9c8d5a43eabb8016a1b0fa2a9548dfe15ff4f232d0bb677913c0c30063dac"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.777395 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerStarted","Data":"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.785751 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerStarted","Data":"01aca9e27d1451bb6122006ca98898ab38ce2caf84632a86b30dbf90281ec871"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.790677 4727 generic.go:334] "Generic (PLEG): container finished" podID="77b2d573-abbb-491f-8939-6c7f79082906" containerID="ff3a5688d5e995bc23e2c566cee352cfc332a17885d01f72318a20848b544e2f" exitCode=0 Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.790738 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rrbt8" event={"ID":"77b2d573-abbb-491f-8939-6c7f79082906","Type":"ContainerDied","Data":"ff3a5688d5e995bc23e2c566cee352cfc332a17885d01f72318a20848b544e2f"} Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.859207 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-559c5978cd-jb7bp" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.859315 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:41:59 crc kubenswrapper[4727]: I0929 10:41:59.910929 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" podStartSLOduration=3.9109089949999998 podStartE2EDuration="3.910908995s" podCreationTimestamp="2025-09-29 10:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:41:59.771619291 +0000 UTC m=+1189.944932653" watchObservedRunningTime="2025-09-29 10:41:59.910908995 +0000 UTC m=+1190.084222357" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.287601 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.387993 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kpq7\" (UniqueName: \"kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7\") pod \"973c7613-3c61-4534-b26e-e1d0ea649bc9\" (UID: \"973c7613-3c61-4534-b26e-e1d0ea649bc9\") " Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.393673 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7" (OuterVolumeSpecName: "kube-api-access-6kpq7") pod "973c7613-3c61-4534-b26e-e1d0ea649bc9" (UID: "973c7613-3c61-4534-b26e-e1d0ea649bc9"). InnerVolumeSpecName "kube-api-access-6kpq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.490771 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kpq7\" (UniqueName: \"kubernetes.io/projected/973c7613-3c61-4534-b26e-e1d0ea649bc9-kube-api-access-6kpq7\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.507553 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.522713 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l6trr" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.594005 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk7q5\" (UniqueName: \"kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5\") pod \"77b2d573-abbb-491f-8939-6c7f79082906\" (UID: \"77b2d573-abbb-491f-8939-6c7f79082906\") " Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.594261 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kddr\" (UniqueName: \"kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr\") pod \"6be32091-cf3a-4931-bcb8-2c867604774e\" (UID: \"6be32091-cf3a-4931-bcb8-2c867604774e\") " Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.625610 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr" (OuterVolumeSpecName: "kube-api-access-2kddr") pod "6be32091-cf3a-4931-bcb8-2c867604774e" (UID: "6be32091-cf3a-4931-bcb8-2c867604774e"). InnerVolumeSpecName "kube-api-access-2kddr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.645485 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5" (OuterVolumeSpecName: "kube-api-access-wk7q5") pod "77b2d573-abbb-491f-8939-6c7f79082906" (UID: "77b2d573-abbb-491f-8939-6c7f79082906"). InnerVolumeSpecName "kube-api-access-wk7q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.699864 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk7q5\" (UniqueName: \"kubernetes.io/projected/77b2d573-abbb-491f-8939-6c7f79082906-kube-api-access-wk7q5\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.699930 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kddr\" (UniqueName: \"kubernetes.io/projected/6be32091-cf3a-4931-bcb8-2c867604774e-kube-api-access-2kddr\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.823744 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rrbt8" event={"ID":"77b2d573-abbb-491f-8939-6c7f79082906","Type":"ContainerDied","Data":"bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.823789 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf78edb53856cd66b6044ad5b75309fdbe27576022332d5354dcbb5e9dd9b754" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.823881 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rrbt8" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.892569 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l6trr" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.894715 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l6trr" event={"ID":"6be32091-cf3a-4931-bcb8-2c867604774e","Type":"ContainerDied","Data":"4db178844fe8130e66b5f25d7cf33b02abfd2f0fec945fbc81de6fd7a35509f3"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.894778 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4db178844fe8130e66b5f25d7cf33b02abfd2f0fec945fbc81de6fd7a35509f3" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.911828 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbvq2" event={"ID":"973c7613-3c61-4534-b26e-e1d0ea649bc9","Type":"ContainerDied","Data":"4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.912057 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4378e36514482c2a8c8b1f6ac89366e7257c6814580657f0136db7b665a51d4a" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.912326 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbvq2" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.934888 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerStarted","Data":"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.952783 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerStarted","Data":"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.952932 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api-log" containerID="cri-o://bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" gracePeriod=30 Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.953186 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api" containerID="cri-o://ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" gracePeriod=30 Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.953206 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.963247 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerStarted","Data":"298323e30d8dd9ad15527d8a5b9384b1d6ed2ac20a464db2e1ccff15b554aabe"} Sep 29 10:42:00 crc kubenswrapper[4727]: I0929 10:42:00.978772 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.978748379 podStartE2EDuration="4.978748379s" podCreationTimestamp="2025-09-29 10:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:42:00.976988104 +0000 UTC m=+1191.150301466" watchObservedRunningTime="2025-09-29 10:42:00.978748379 +0000 UTC m=+1191.152061741" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.682142 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.750446 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.750557 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.750601 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nz2d\" (UniqueName: \"kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.750667 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.750662 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.752501 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.752794 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.752926 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data\") pod \"a26a02c4-5b59-41fc-b72d-972016013b25\" (UID: \"a26a02c4-5b59-41fc-b72d-972016013b25\") " Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.753467 4727 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a26a02c4-5b59-41fc-b72d-972016013b25-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.755061 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs" (OuterVolumeSpecName: "logs") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.756043 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d" (OuterVolumeSpecName: "kube-api-access-5nz2d") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "kube-api-access-5nz2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.757292 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts" (OuterVolumeSpecName: "scripts") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.759413 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.827726 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data" (OuterVolumeSpecName: "config-data") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.842462 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a26a02c4-5b59-41fc-b72d-972016013b25" (UID: "a26a02c4-5b59-41fc-b72d-972016013b25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.855708 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.856044 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.856055 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nz2d\" (UniqueName: \"kubernetes.io/projected/a26a02c4-5b59-41fc-b72d-972016013b25-kube-api-access-5nz2d\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.856066 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.856075 4727 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a26a02c4-5b59-41fc-b72d-972016013b25-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.856085 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a26a02c4-5b59-41fc-b72d-972016013b25-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973563 4727 generic.go:334] "Generic (PLEG): container finished" podID="a26a02c4-5b59-41fc-b72d-972016013b25" containerID="ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" exitCode=0 Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973680 4727 generic.go:334] "Generic (PLEG): container finished" podID="a26a02c4-5b59-41fc-b72d-972016013b25" containerID="bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" exitCode=143 Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973660 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973678 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerDied","Data":"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd"} Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973848 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerDied","Data":"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d"} Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973866 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a26a02c4-5b59-41fc-b72d-972016013b25","Type":"ContainerDied","Data":"962101826d72f2be1e526162a08e12be220d164a1232c2acd9d71e206d770833"} Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.973883 4727 scope.go:117] "RemoveContainer" containerID="ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.980282 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerStarted","Data":"45ddc7ece121325b027ad3c2184aedb91824e770b1258a3d5fa6384410c7e782"} Sep 29 10:42:01 crc kubenswrapper[4727]: I0929 10:42:01.989385 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerStarted","Data":"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b"} Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.004638 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.007041 4727 scope.go:117] "RemoveContainer" containerID="bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.012429 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025094 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.025481 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973c7613-3c61-4534-b26e-e1d0ea649bc9" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025497 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="973c7613-3c61-4534-b26e-e1d0ea649bc9" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.025528 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025540 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.025555 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be32091-cf3a-4931-bcb8-2c867604774e" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025561 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be32091-cf3a-4931-bcb8-2c867604774e" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.025575 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b2d573-abbb-491f-8939-6c7f79082906" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025581 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b2d573-abbb-491f-8939-6c7f79082906" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.025591 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api-log" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025597 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api-log" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025785 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be32091-cf3a-4931-bcb8-2c867604774e" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025799 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b2d573-abbb-491f-8939-6c7f79082906" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025815 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025830 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="973c7613-3c61-4534-b26e-e1d0ea649bc9" containerName="mariadb-database-create" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.025840 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" containerName="cinder-api-log" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.030415 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.035572 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.035681 4727 scope.go:117] "RemoveContainer" containerID="ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.035725 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.035845 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.039021 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.91157145 podStartE2EDuration="6.039007588s" podCreationTimestamp="2025-09-29 10:41:56 +0000 UTC" firstStartedPulling="2025-09-29 10:41:58.159675418 +0000 UTC m=+1188.332988780" lastFinishedPulling="2025-09-29 10:41:59.287111556 +0000 UTC m=+1189.460424918" observedRunningTime="2025-09-29 10:42:02.016821877 +0000 UTC m=+1192.190135239" watchObservedRunningTime="2025-09-29 10:42:02.039007588 +0000 UTC m=+1192.212320940" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.039536 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd\": container with ID starting with ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd not found: ID does not exist" containerID="ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.039580 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd"} err="failed to get container status \"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd\": rpc error: code = NotFound desc = could not find container \"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd\": container with ID starting with ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd not found: ID does not exist" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.039608 4727 scope.go:117] "RemoveContainer" containerID="bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" Sep 29 10:42:02 crc kubenswrapper[4727]: E0929 10:42:02.039932 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d\": container with ID starting with bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d not found: ID does not exist" containerID="bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.039960 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d"} err="failed to get container status \"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d\": rpc error: code = NotFound desc = could not find container \"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d\": container with ID starting with bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d not found: ID does not exist" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.039983 4727 scope.go:117] "RemoveContainer" containerID="ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.040202 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd"} err="failed to get container status \"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd\": rpc error: code = NotFound desc = could not find container \"ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd\": container with ID starting with ee9530f57781931acc0f00b2af5b0f16f2b31ae3dfb4db4156097b1955df5bdd not found: ID does not exist" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.040223 4727 scope.go:117] "RemoveContainer" containerID="bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.040428 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d"} err="failed to get container status \"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d\": rpc error: code = NotFound desc = could not find container \"bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d\": container with ID starting with bd8f849be3968db19dd1e62be72f03bbacc3a4f329852e982bc64953f22a5e2d not found: ID does not exist" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.055388 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162783 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data-custom\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162841 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0248c486-ae28-4a7c-96f6-e97cde83ae7b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162881 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162910 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0248c486-ae28-4a7c-96f6-e97cde83ae7b-logs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162935 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.162988 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.163017 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s6qr\" (UniqueName: \"kubernetes.io/projected/0248c486-ae28-4a7c-96f6-e97cde83ae7b-kube-api-access-8s6qr\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.163461 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.163728 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-scripts\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265557 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-scripts\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265656 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data-custom\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265679 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0248c486-ae28-4a7c-96f6-e97cde83ae7b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265704 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265722 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0248c486-ae28-4a7c-96f6-e97cde83ae7b-logs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265738 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265766 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265792 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s6qr\" (UniqueName: \"kubernetes.io/projected/0248c486-ae28-4a7c-96f6-e97cde83ae7b-kube-api-access-8s6qr\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.265822 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.266752 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0248c486-ae28-4a7c-96f6-e97cde83ae7b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.267303 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0248c486-ae28-4a7c-96f6-e97cde83ae7b-logs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.271522 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-scripts\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.272123 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.272702 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data-custom\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.273177 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.279148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.279229 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0248c486-ae28-4a7c-96f6-e97cde83ae7b-config-data\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.287839 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s6qr\" (UniqueName: \"kubernetes.io/projected/0248c486-ae28-4a7c-96f6-e97cde83ae7b-kube-api-access-8s6qr\") pod \"cinder-api-0\" (UID: \"0248c486-ae28-4a7c-96f6-e97cde83ae7b\") " pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.316088 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.349115 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 29 10:42:02 crc kubenswrapper[4727]: I0929 10:42:02.796935 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.000591 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-feac-account-create-j9dr8"] Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.001742 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.006381 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.021083 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-feac-account-create-j9dr8"] Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.028109 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0248c486-ae28-4a7c-96f6-e97cde83ae7b","Type":"ContainerStarted","Data":"6bb5b91d1591ed710c820bd766b2829fefa31a7dd961027a24709d4e78d94097"} Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.084018 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmm7g\" (UniqueName: \"kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g\") pod \"nova-api-feac-account-create-j9dr8\" (UID: \"6f04c134-ab93-4a72-9e26-eee995532432\") " pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.133760 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a26a02c4-5b59-41fc-b72d-972016013b25" path="/var/lib/kubelet/pods/a26a02c4-5b59-41fc-b72d-972016013b25/volumes" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.185971 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmm7g\" (UniqueName: \"kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g\") pod \"nova-api-feac-account-create-j9dr8\" (UID: \"6f04c134-ab93-4a72-9e26-eee995532432\") " pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.209884 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmm7g\" (UniqueName: \"kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g\") pod \"nova-api-feac-account-create-j9dr8\" (UID: \"6f04c134-ab93-4a72-9e26-eee995532432\") " pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:03 crc kubenswrapper[4727]: I0929 10:42:03.326636 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:04 crc kubenswrapper[4727]: I0929 10:42:04.065993 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0248c486-ae28-4a7c-96f6-e97cde83ae7b","Type":"ContainerStarted","Data":"bc14d3e0714ab8c6ef192717a3b49668cb754c6062f069f8518196e6a66d7d8e"} Sep 29 10:42:04 crc kubenswrapper[4727]: I0929 10:42:04.479172 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-feac-account-create-j9dr8"] Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.083052 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerStarted","Data":"8f043070636bd3d5d82f76cb8f0a259e4c5c2f308884e5e06c22e46fe87656f6"} Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.083481 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.091312 4727 generic.go:334] "Generic (PLEG): container finished" podID="6f04c134-ab93-4a72-9e26-eee995532432" containerID="94e3bf00423422c8c394a3220719710323321caddfb62bd0fb53061271bd7f5a" exitCode=0 Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.091385 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-feac-account-create-j9dr8" event={"ID":"6f04c134-ab93-4a72-9e26-eee995532432","Type":"ContainerDied","Data":"94e3bf00423422c8c394a3220719710323321caddfb62bd0fb53061271bd7f5a"} Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.091441 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-feac-account-create-j9dr8" event={"ID":"6f04c134-ab93-4a72-9e26-eee995532432","Type":"ContainerStarted","Data":"e2823266b6a33a0d96e169d1ec3c6b7ecc1a8573b952f4e9ee96f48044af8688"} Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.097081 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0248c486-ae28-4a7c-96f6-e97cde83ae7b","Type":"ContainerStarted","Data":"ad6a6d798cb89f2315f46d346daa15a7abbc0a9a506f7434b5dd6dec03529a6a"} Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.097216 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.102222 4727 generic.go:334] "Generic (PLEG): container finished" podID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerID="4cd6244f38b2627ab0c262de29eecfcff1b39f6d175990a0f329ccd2991f7f30" exitCode=137 Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.102280 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerDied","Data":"4cd6244f38b2627ab0c262de29eecfcff1b39f6d175990a0f329ccd2991f7f30"} Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.107468 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.239898472 podStartE2EDuration="15.107452004s" podCreationTimestamp="2025-09-29 10:41:50 +0000 UTC" firstStartedPulling="2025-09-29 10:41:57.112476965 +0000 UTC m=+1187.285790327" lastFinishedPulling="2025-09-29 10:42:03.980030507 +0000 UTC m=+1194.153343859" observedRunningTime="2025-09-29 10:42:05.104869777 +0000 UTC m=+1195.278183139" watchObservedRunningTime="2025-09-29 10:42:05.107452004 +0000 UTC m=+1195.280765366" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.176401 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.176383857 podStartE2EDuration="3.176383857s" podCreationTimestamp="2025-09-29 10:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:42:05.173255977 +0000 UTC m=+1195.346569339" watchObservedRunningTime="2025-09-29 10:42:05.176383857 +0000 UTC m=+1195.349697219" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.372375 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424434 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424521 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424568 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktrz5\" (UniqueName: \"kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424607 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424733 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424753 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.424808 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data\") pod \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\" (UID: \"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c\") " Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.432001 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs" (OuterVolumeSpecName: "logs") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.432655 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5" (OuterVolumeSpecName: "kube-api-access-ktrz5") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "kube-api-access-ktrz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.449176 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.471907 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data" (OuterVolumeSpecName: "config-data") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.476078 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts" (OuterVolumeSpecName: "scripts") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.497683 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.511878 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" (UID: "a9fd37e0-aad8-4319-a10f-4b27aaa56a2c"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527533 4727 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527572 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527581 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527591 4727 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527602 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527611 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktrz5\" (UniqueName: \"kubernetes.io/projected/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-kube-api-access-ktrz5\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:05 crc kubenswrapper[4727]: I0929 10:42:05.527620 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.114440 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-559c5978cd-jb7bp" event={"ID":"a9fd37e0-aad8-4319-a10f-4b27aaa56a2c","Type":"ContainerDied","Data":"481952bf105362fb6dce800e1009a43adcb4406f1e2796fa433d14461722dab4"} Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.114498 4727 scope.go:117] "RemoveContainer" containerID="d4e2464b9ae71a314ef444db324581c200bec588ddbb7b07b1c93c3d46fa4633" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.114623 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-559c5978cd-jb7bp" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.151014 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.158888 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-559c5978cd-jb7bp"] Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.325947 4727 scope.go:117] "RemoveContainer" containerID="4cd6244f38b2627ab0c262de29eecfcff1b39f6d175990a0f329ccd2991f7f30" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.565822 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.652224 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmm7g\" (UniqueName: \"kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g\") pod \"6f04c134-ab93-4a72-9e26-eee995532432\" (UID: \"6f04c134-ab93-4a72-9e26-eee995532432\") " Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.656433 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g" (OuterVolumeSpecName: "kube-api-access-fmm7g") pod "6f04c134-ab93-4a72-9e26-eee995532432" (UID: "6f04c134-ab93-4a72-9e26-eee995532432"). InnerVolumeSpecName "kube-api-access-fmm7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:06 crc kubenswrapper[4727]: I0929 10:42:06.754882 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmm7g\" (UniqueName: \"kubernetes.io/projected/6f04c134-ab93-4a72-9e26-eee995532432-kube-api-access-fmm7g\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.121593 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" path="/var/lib/kubelet/pods/a9fd37e0-aad8-4319-a10f-4b27aaa56a2c/volumes" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.128257 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-feac-account-create-j9dr8" event={"ID":"6f04c134-ab93-4a72-9e26-eee995532432","Type":"ContainerDied","Data":"e2823266b6a33a0d96e169d1ec3c6b7ecc1a8573b952f4e9ee96f48044af8688"} Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.128597 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2823266b6a33a0d96e169d1ec3c6b7ecc1a8573b952f4e9ee96f48044af8688" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.128326 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-feac-account-create-j9dr8" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.195682 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.264261 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.264769 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="dnsmasq-dns" containerID="cri-o://66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484" gracePeriod=10 Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.402532 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.402771 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-log" containerID="cri-o://3809597929b99fccd1f06ad19ec56e9c734668bea3069883bac20cdb27913ae7" gracePeriod=30 Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.403282 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-httpd" containerID="cri-o://a0a6dfca2b0543c83afb7c1f48c5eb590f228bdebf1a3522a00edaa1e3a0a59f" gracePeriod=30 Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.631804 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.719838 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.887594 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.980637 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.980700 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.980798 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.980933 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ngzp\" (UniqueName: \"kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.980964 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.981018 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb\") pod \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\" (UID: \"17498e09-aa36-4ba6-8942-f8c18bdc8f78\") " Sep 29 10:42:07 crc kubenswrapper[4727]: I0929 10:42:07.987032 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp" (OuterVolumeSpecName: "kube-api-access-9ngzp") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "kube-api-access-9ngzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.032148 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.032989 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.035399 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config" (OuterVolumeSpecName: "config") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.047881 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.061771 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17498e09-aa36-4ba6-8942-f8c18bdc8f78" (UID: "17498e09-aa36-4ba6-8942-f8c18bdc8f78"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082683 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082715 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082726 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082735 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082745 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ngzp\" (UniqueName: \"kubernetes.io/projected/17498e09-aa36-4ba6-8942-f8c18bdc8f78-kube-api-access-9ngzp\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.082753 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17498e09-aa36-4ba6-8942-f8c18bdc8f78-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.139284 4727 generic.go:334] "Generic (PLEG): container finished" podID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerID="3809597929b99fccd1f06ad19ec56e9c734668bea3069883bac20cdb27913ae7" exitCode=143 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.139404 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerDied","Data":"3809597929b99fccd1f06ad19ec56e9c734668bea3069883bac20cdb27913ae7"} Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141473 4727 generic.go:334] "Generic (PLEG): container finished" podID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerID="66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484" exitCode=0 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141669 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" event={"ID":"17498e09-aa36-4ba6-8942-f8c18bdc8f78","Type":"ContainerDied","Data":"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484"} Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141729 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="cinder-scheduler" containerID="cri-o://a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141750 4727 scope.go:117] "RemoveContainer" containerID="66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141761 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141841 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="probe" containerID="cri-o://c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.141733 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" event={"ID":"17498e09-aa36-4ba6-8942-f8c18bdc8f78","Type":"ContainerDied","Data":"690ed332b4a63dd6b7e150e456e15e288b04fa2e5861efe46e6b2e4ae6194fac"} Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.172300 4727 scope.go:117] "RemoveContainer" containerID="ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.195564 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.205984 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64667c4f57-m4xhc"] Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.213664 4727 scope.go:117] "RemoveContainer" containerID="66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484" Sep 29 10:42:08 crc kubenswrapper[4727]: E0929 10:42:08.214235 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484\": container with ID starting with 66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484 not found: ID does not exist" containerID="66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.214272 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484"} err="failed to get container status \"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484\": rpc error: code = NotFound desc = could not find container \"66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484\": container with ID starting with 66cd9f3acb8cebdd31ca21dccf62a8b609c25d650896e277bcb0641c15aea484 not found: ID does not exist" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.214299 4727 scope.go:117] "RemoveContainer" containerID="ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37" Sep 29 10:42:08 crc kubenswrapper[4727]: E0929 10:42:08.214593 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37\": container with ID starting with ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37 not found: ID does not exist" containerID="ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.214625 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37"} err="failed to get container status \"ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37\": rpc error: code = NotFound desc = could not find container \"ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37\": container with ID starting with ff9192bade285815a7cd53c3b6dc7efa0208c87e8a5e790f16fba20e3b443f37 not found: ID does not exist" Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.386065 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.386394 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-central-agent" containerID="cri-o://01aca9e27d1451bb6122006ca98898ab38ce2caf84632a86b30dbf90281ec871" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.386540 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="proxy-httpd" containerID="cri-o://8f043070636bd3d5d82f76cb8f0a259e4c5c2f308884e5e06c22e46fe87656f6" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.386583 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="sg-core" containerID="cri-o://45ddc7ece121325b027ad3c2184aedb91824e770b1258a3d5fa6384410c7e782" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.386618 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-notification-agent" containerID="cri-o://298323e30d8dd9ad15527d8a5b9384b1d6ed2ac20a464db2e1ccff15b554aabe" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.537278 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.540570 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-log" containerID="cri-o://8d8aa7de856cc6bf1639b3a983d2c3f34670e982a11e9d4d806252c2094a5c17" gracePeriod=30 Sep 29 10:42:08 crc kubenswrapper[4727]: I0929 10:42:08.540771 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-httpd" containerID="cri-o://ae0b44851e2f70479e7adc332a042a7f624d3c636a668dba6b88fa26d8703144" gracePeriod=30 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.121635 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" path="/var/lib/kubelet/pods/17498e09-aa36-4ba6-8942-f8c18bdc8f78/volumes" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.159447 4727 generic.go:334] "Generic (PLEG): container finished" podID="f00fce5b-a355-42e2-94c7-63810205536f" containerID="8d8aa7de856cc6bf1639b3a983d2c3f34670e982a11e9d4d806252c2094a5c17" exitCode=143 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.159527 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerDied","Data":"8d8aa7de856cc6bf1639b3a983d2c3f34670e982a11e9d4d806252c2094a5c17"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.162756 4727 generic.go:334] "Generic (PLEG): container finished" podID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerID="a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d" exitCode=0 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.162800 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerDied","Data":"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169362 4727 generic.go:334] "Generic (PLEG): container finished" podID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerID="8f043070636bd3d5d82f76cb8f0a259e4c5c2f308884e5e06c22e46fe87656f6" exitCode=0 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169405 4727 generic.go:334] "Generic (PLEG): container finished" podID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerID="45ddc7ece121325b027ad3c2184aedb91824e770b1258a3d5fa6384410c7e782" exitCode=2 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169415 4727 generic.go:334] "Generic (PLEG): container finished" podID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerID="298323e30d8dd9ad15527d8a5b9384b1d6ed2ac20a464db2e1ccff15b554aabe" exitCode=0 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169424 4727 generic.go:334] "Generic (PLEG): container finished" podID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerID="01aca9e27d1451bb6122006ca98898ab38ce2caf84632a86b30dbf90281ec871" exitCode=0 Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169446 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerDied","Data":"8f043070636bd3d5d82f76cb8f0a259e4c5c2f308884e5e06c22e46fe87656f6"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169474 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerDied","Data":"45ddc7ece121325b027ad3c2184aedb91824e770b1258a3d5fa6384410c7e782"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169486 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerDied","Data":"298323e30d8dd9ad15527d8a5b9384b1d6ed2ac20a464db2e1ccff15b554aabe"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.169497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerDied","Data":"01aca9e27d1451bb6122006ca98898ab38ce2caf84632a86b30dbf90281ec871"} Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.212332 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.328637 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.328741 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.328815 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.328846 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.328924 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.329033 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjzs2\" (UniqueName: \"kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.329123 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml\") pod \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\" (UID: \"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.343461 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.344025 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.359674 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts" (OuterVolumeSpecName: "scripts") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.411623 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.414993 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2" (OuterVolumeSpecName: "kube-api-access-vjzs2") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "kube-api-access-vjzs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.433191 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjzs2\" (UniqueName: \"kubernetes.io/projected/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-kube-api-access-vjzs2\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.433217 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.433228 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.433236 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.433245 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.598193 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.639959 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.640469 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data" (OuterVolumeSpecName: "config-data") pod "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" (UID: "5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.708094 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.741986 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.843574 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.843639 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.843674 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.844203 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.844323 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9866\" (UniqueName: \"kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.844756 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id\") pod \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\" (UID: \"bf74dfd8-697e-477a-82fb-62b76c4a18a4\") " Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.844836 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.845416 4727 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf74dfd8-697e-477a-82fb-62b76c4a18a4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.847572 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts" (OuterVolumeSpecName: "scripts") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.848145 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866" (OuterVolumeSpecName: "kube-api-access-s9866") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "kube-api-access-s9866". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.851497 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.906300 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.947734 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9866\" (UniqueName: \"kubernetes.io/projected/bf74dfd8-697e-477a-82fb-62b76c4a18a4-kube-api-access-s9866\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.947781 4727 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.947793 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.947807 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:09 crc kubenswrapper[4727]: I0929 10:42:09.990817 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data" (OuterVolumeSpecName: "config-data") pod "bf74dfd8-697e-477a-82fb-62b76c4a18a4" (UID: "bf74dfd8-697e-477a-82fb-62b76c4a18a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.049332 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf74dfd8-697e-477a-82fb-62b76c4a18a4-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.180531 4727 generic.go:334] "Generic (PLEG): container finished" podID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerID="c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b" exitCode=0 Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.180592 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerDied","Data":"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b"} Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.180619 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf74dfd8-697e-477a-82fb-62b76c4a18a4","Type":"ContainerDied","Data":"2ab5a70892d899c39539ea4622e709243326a077029500f69450f8ddb9ac16eb"} Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.180635 4727 scope.go:117] "RemoveContainer" containerID="c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.180735 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.191357 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10","Type":"ContainerDied","Data":"426c653ea90f98db398dd2c1691b49dfc74b9ecb04f9609ea64271a6d8f79ea4"} Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.191473 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.216231 4727 scope.go:117] "RemoveContainer" containerID="a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.219771 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.242219 4727 scope.go:117] "RemoveContainer" containerID="c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.242896 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b\": container with ID starting with c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b not found: ID does not exist" containerID="c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.243012 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b"} err="failed to get container status \"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b\": rpc error: code = NotFound desc = could not find container \"c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b\": container with ID starting with c996840bb33a76901e3ec3a18c9c98240516a0ab4dc79ed195f5cd3fd5ebd45b not found: ID does not exist" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.243112 4727 scope.go:117] "RemoveContainer" containerID="a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.243469 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d\": container with ID starting with a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d not found: ID does not exist" containerID="a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.243581 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d"} err="failed to get container status \"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d\": rpc error: code = NotFound desc = could not find container \"a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d\": container with ID starting with a04bfd510a2487cf93caed9308f3c9ae0963087f71070be122060c3bd75a3b9d not found: ID does not exist" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.243665 4727 scope.go:117] "RemoveContainer" containerID="8f043070636bd3d5d82f76cb8f0a259e4c5c2f308884e5e06c22e46fe87656f6" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.243850 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254090 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254519 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254540 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254554 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-central-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254562 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-central-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254580 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="cinder-scheduler" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254588 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="cinder-scheduler" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254607 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="proxy-httpd" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254615 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="proxy-httpd" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254625 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="dnsmasq-dns" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254632 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="dnsmasq-dns" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254644 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-notification-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254651 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-notification-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254671 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="sg-core" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254677 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="sg-core" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254687 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="probe" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254693 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="probe" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254701 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f04c134-ab93-4a72-9e26-eee995532432" containerName="mariadb-account-create" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254708 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f04c134-ab93-4a72-9e26-eee995532432" containerName="mariadb-account-create" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254718 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="init" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254724 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="init" Sep 29 10:42:10 crc kubenswrapper[4727]: E0929 10:42:10.254734 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon-log" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254741 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon-log" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254895 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="sg-core" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254910 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254920 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f04c134-ab93-4a72-9e26-eee995532432" containerName="mariadb-account-create" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254932 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="probe" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254942 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-central-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254953 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="dnsmasq-dns" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254964 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" containerName="cinder-scheduler" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254972 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fd37e0-aad8-4319-a10f-4b27aaa56a2c" containerName="horizon-log" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.254988 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="ceilometer-notification-agent" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.255001 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" containerName="proxy-httpd" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.265503 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.265622 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.268498 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.268739 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.268842 4727 scope.go:117] "RemoveContainer" containerID="45ddc7ece121325b027ad3c2184aedb91824e770b1258a3d5fa6384410c7e782" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.276750 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.283252 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.285553 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.289489 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.289578 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.303295 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.327849 4727 scope.go:117] "RemoveContainer" containerID="298323e30d8dd9ad15527d8a5b9384b1d6ed2ac20a464db2e1ccff15b554aabe" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.350852 4727 scope.go:117] "RemoveContainer" containerID="01aca9e27d1451bb6122006ca98898ab38ce2caf84632a86b30dbf90281ec871" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354691 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354753 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354779 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354831 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354868 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae9bf7b1-b3a8-4b34-8930-3751c0230705-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354887 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354912 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354937 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.354991 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.355031 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj26l\" (UniqueName: \"kubernetes.io/projected/ae9bf7b1-b3a8-4b34-8930-3751c0230705-kube-api-access-lj26l\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.355056 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv2kc\" (UniqueName: \"kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.355075 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.355107 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457065 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae9bf7b1-b3a8-4b34-8930-3751c0230705-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457105 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457135 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457166 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457192 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae9bf7b1-b3a8-4b34-8930-3751c0230705-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457221 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457274 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj26l\" (UniqueName: \"kubernetes.io/projected/ae9bf7b1-b3a8-4b34-8930-3751c0230705-kube-api-access-lj26l\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457292 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457307 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv2kc\" (UniqueName: \"kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457348 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457371 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457402 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457418 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.457441 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.458085 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.458290 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.461664 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.461848 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.464374 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.464677 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.468459 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.474721 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv2kc\" (UniqueName: \"kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.475144 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.476611 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9bf7b1-b3a8-4b34-8930-3751c0230705-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.476845 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data\") pod \"ceilometer-0\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.479257 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj26l\" (UniqueName: \"kubernetes.io/projected/ae9bf7b1-b3a8-4b34-8930-3751c0230705-kube-api-access-lj26l\") pod \"cinder-scheduler-0\" (UID: \"ae9bf7b1-b3a8-4b34-8930-3751c0230705\") " pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.601779 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.622238 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:10 crc kubenswrapper[4727]: I0929 10:42:10.664287 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.134287 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10" path="/var/lib/kubelet/pods/5ce03d5c-680e-4f2a-8cfc-fc01c3cd8c10/volumes" Sep 29 10:42:11 crc kubenswrapper[4727]: E0929 10:42:11.135366 4727 info.go:109] Failed to get network devices: open /sys/class/net/a87998dcd3dd24a/address: no such file or directory Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.138608 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf74dfd8-697e-477a-82fb-62b76c4a18a4" path="/var/lib/kubelet/pods/bf74dfd8-697e-477a-82fb-62b76c4a18a4/volumes" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.191930 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.206736 4727 generic.go:334] "Generic (PLEG): container finished" podID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerID="a0a6dfca2b0543c83afb7c1f48c5eb590f228bdebf1a3522a00edaa1e3a0a59f" exitCode=0 Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.206806 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerDied","Data":"a0a6dfca2b0543c83afb7c1f48c5eb590f228bdebf1a3522a00edaa1e3a0a59f"} Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.209958 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerStarted","Data":"7f75a3d2efb9a6d31cfab993e62838f21bcc2577c88dbff80a08db86e1d5c620"} Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.301378 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.302105 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 29 10:42:11 crc kubenswrapper[4727]: W0929 10:42:11.304879 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae9bf7b1_b3a8_4b34_8930_3751c0230705.slice/crio-2aaf27eae6a3cf8434d09489f62295c0bda4be2da0f352ecd5d4e251b5994177 WatchSource:0}: Error finding container 2aaf27eae6a3cf8434d09489f62295c0bda4be2da0f352ecd5d4e251b5994177: Status 404 returned error can't find the container with id 2aaf27eae6a3cf8434d09489f62295c0bda4be2da0f352ecd5d4e251b5994177 Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.396730 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.398163 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.397218 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.398585 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs" (OuterVolumeSpecName: "logs") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.398887 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.399202 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.399785 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.400086 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2npp8\" (UniqueName: \"kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.400256 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.400472 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\" (UID: \"f9249e4e-686b-440c-9cc4-c4f6c1615de2\") " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.404172 4727 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.404329 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9249e4e-686b-440c-9cc4-c4f6c1615de2-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.405862 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts" (OuterVolumeSpecName: "scripts") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.405911 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8" (OuterVolumeSpecName: "kube-api-access-2npp8") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "kube-api-access-2npp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.407082 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.438937 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.470512 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data" (OuterVolumeSpecName: "config-data") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.472018 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f9249e4e-686b-440c-9cc4-c4f6c1615de2" (UID: "f9249e4e-686b-440c-9cc4-c4f6c1615de2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506478 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506530 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506550 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506562 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506576 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2npp8\" (UniqueName: \"kubernetes.io/projected/f9249e4e-686b-440c-9cc4-c4f6c1615de2-kube-api-access-2npp8\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.506589 4727 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9249e4e-686b-440c-9cc4-c4f6c1615de2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.544537 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Sep 29 10:42:11 crc kubenswrapper[4727]: I0929 10:42:11.608695 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.259017 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae9bf7b1-b3a8-4b34-8930-3751c0230705","Type":"ContainerStarted","Data":"d8f7904a06aa45098ccad87771e38f9faeaa3b68d16ea7f55a5df8faa409ac28"} Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.259427 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae9bf7b1-b3a8-4b34-8930-3751c0230705","Type":"ContainerStarted","Data":"2aaf27eae6a3cf8434d09489f62295c0bda4be2da0f352ecd5d4e251b5994177"} Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.284771 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.284792 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9249e4e-686b-440c-9cc4-c4f6c1615de2","Type":"ContainerDied","Data":"a87998dcd3dd24a025ca346bbd4cbdfc812e118d5e2c887dba3a2af244d265f2"} Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.284855 4727 scope.go:117] "RemoveContainer" containerID="a0a6dfca2b0543c83afb7c1f48c5eb590f228bdebf1a3522a00edaa1e3a0a59f" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.293649 4727 generic.go:334] "Generic (PLEG): container finished" podID="f00fce5b-a355-42e2-94c7-63810205536f" containerID="ae0b44851e2f70479e7adc332a042a7f624d3c636a668dba6b88fa26d8703144" exitCode=0 Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.293971 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerDied","Data":"ae0b44851e2f70479e7adc332a042a7f624d3c636a668dba6b88fa26d8703144"} Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.338882 4727 scope.go:117] "RemoveContainer" containerID="3809597929b99fccd1f06ad19ec56e9c734668bea3069883bac20cdb27913ae7" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.386327 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.437477 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.442897 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.485114 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:12 crc kubenswrapper[4727]: E0929 10:42:12.486297 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486319 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: E0929 10:42:12.486393 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486402 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: E0929 10:42:12.486417 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486423 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: E0929 10:42:12.486445 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486450 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486643 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486657 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486670 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00fce5b-a355-42e2-94c7-63810205536f" containerName="glance-httpd" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.486680 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" containerName="glance-log" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.487828 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.493528 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.494392 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536018 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536107 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536129 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536160 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536224 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536281 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gqw7\" (UniqueName: \"kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536324 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.536365 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs\") pod \"f00fce5b-a355-42e2-94c7-63810205536f\" (UID: \"f00fce5b-a355-42e2-94c7-63810205536f\") " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.544278 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.544913 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs" (OuterVolumeSpecName: "logs") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.545450 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.550697 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts" (OuterVolumeSpecName: "scripts") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.560876 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.573385 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7" (OuterVolumeSpecName: "kube-api-access-9gqw7") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "kube-api-access-9gqw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644185 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644233 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644252 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644320 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644357 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644407 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-logs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644433 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644484 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9qdz\" (UniqueName: \"kubernetes.io/projected/56bbd245-faf2-4c21-b4c9-172a501374fd-kube-api-access-p9qdz\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644534 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gqw7\" (UniqueName: \"kubernetes.io/projected/f00fce5b-a355-42e2-94c7-63810205536f-kube-api-access-9gqw7\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644554 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644563 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644573 4727 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.644581 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00fce5b-a355-42e2-94c7-63810205536f-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.646231 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.665014 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.670870 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.692677 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64667c4f57-m4xhc" podUID="17498e09-aa36-4ba6-8942-f8c18bdc8f78" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.165:5353: i/o timeout" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.705529 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data" (OuterVolumeSpecName: "config-data") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.705903 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f00fce5b-a355-42e2-94c7-63810205536f" (UID: "f00fce5b-a355-42e2-94c7-63810205536f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.747696 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.747757 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.747907 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-logs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.747975 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748075 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9qdz\" (UniqueName: \"kubernetes.io/projected/56bbd245-faf2-4c21-b4c9-172a501374fd-kube-api-access-p9qdz\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748134 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748168 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748189 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748354 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748375 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748388 4727 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.748403 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00fce5b-a355-42e2-94c7-63810205536f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.750900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-logs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.751463 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56bbd245-faf2-4c21-b4c9-172a501374fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.751964 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.756916 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.758525 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.758545 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.772481 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56bbd245-faf2-4c21-b4c9-172a501374fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.780118 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9qdz\" (UniqueName: \"kubernetes.io/projected/56bbd245-faf2-4c21-b4c9-172a501374fd-kube-api-access-p9qdz\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:12 crc kubenswrapper[4727]: I0929 10:42:12.829518 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"56bbd245-faf2-4c21-b4c9-172a501374fd\") " pod="openstack/glance-default-external-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.113643 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.129736 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9249e4e-686b-440c-9cc4-c4f6c1615de2" path="/var/lib/kubelet/pods/f9249e4e-686b-440c-9cc4-c4f6c1615de2/volumes" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.199415 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ad4a-account-create-blgkd"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.200634 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.206037 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.211938 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ad4a-account-create-blgkd"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.260268 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k79ff\" (UniqueName: \"kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff\") pod \"nova-cell0-ad4a-account-create-blgkd\" (UID: \"63978328-20b2-46a4-a423-b9c15c0c03cd\") " pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.330655 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae9bf7b1-b3a8-4b34-8930-3751c0230705","Type":"ContainerStarted","Data":"5b3e49603e69bb58676559d71e61debeccb196fac869a03b11b22e18893bf139"} Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.356155 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f00fce5b-a355-42e2-94c7-63810205536f","Type":"ContainerDied","Data":"89acd8e1f6013b6d434f36580975c6c8da4d309201dc1dba406964b53d22ca93"} Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.358304 4727 scope.go:117] "RemoveContainer" containerID="ae0b44851e2f70479e7adc332a042a7f624d3c636a668dba6b88fa26d8703144" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.358565 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.363699 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k79ff\" (UniqueName: \"kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff\") pod \"nova-cell0-ad4a-account-create-blgkd\" (UID: \"63978328-20b2-46a4-a423-b9c15c0c03cd\") " pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.370456 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.37043621 podStartE2EDuration="3.37043621s" podCreationTimestamp="2025-09-29 10:42:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:42:13.355894245 +0000 UTC m=+1203.529207627" watchObservedRunningTime="2025-09-29 10:42:13.37043621 +0000 UTC m=+1203.543749572" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.381418 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerStarted","Data":"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da"} Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.392093 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k79ff\" (UniqueName: \"kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff\") pod \"nova-cell0-ad4a-account-create-blgkd\" (UID: \"63978328-20b2-46a4-a423-b9c15c0c03cd\") " pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.398654 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.435913 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.468355 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f241-account-create-rk5l6"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.469743 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.477549 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.488639 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.490274 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.494947 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.495154 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.505518 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f241-account-create-rk5l6"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.535489 4727 scope.go:117] "RemoveContainer" containerID="8d8aa7de856cc6bf1639b3a983d2c3f34670e982a11e9d4d806252c2094a5c17" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.537251 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.551925 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578280 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578328 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlhpq\" (UniqueName: \"kubernetes.io/projected/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-kube-api-access-xlhpq\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578387 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578403 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578469 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578490 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578505 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xjw9\" (UniqueName: \"kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9\") pod \"nova-cell1-f241-account-create-rk5l6\" (UID: \"2eedcefb-dda0-4045-abbd-611299c470aa\") " pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578545 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.578575 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-logs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.679972 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680304 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xjw9\" (UniqueName: \"kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9\") pod \"nova-cell1-f241-account-create-rk5l6\" (UID: \"2eedcefb-dda0-4045-abbd-611299c470aa\") " pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680324 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680401 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680440 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-logs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680480 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680500 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlhpq\" (UniqueName: \"kubernetes.io/projected/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-kube-api-access-xlhpq\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680537 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.680551 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.681535 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.682424 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-logs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.685374 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.694364 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.698179 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.705223 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.708148 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xjw9\" (UniqueName: \"kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9\") pod \"nova-cell1-f241-account-create-rk5l6\" (UID: \"2eedcefb-dda0-4045-abbd-611299c470aa\") " pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.709045 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.782325 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlhpq\" (UniqueName: \"kubernetes.io/projected/b1ab9d58-dea7-47fc-a259-5b9b8f7a7230-kube-api-access-xlhpq\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.813318 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230\") " pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.837971 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.848149 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:13 crc kubenswrapper[4727]: I0929 10:42:13.957670 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.213566 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ad4a-account-create-blgkd"] Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.405827 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad4a-account-create-blgkd" event={"ID":"63978328-20b2-46a4-a423-b9c15c0c03cd","Type":"ContainerStarted","Data":"176d60411f9e00549cb3f23d70594fa3dc936bc83ff06c1e623a9854a29c429d"} Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.414595 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56bbd245-faf2-4c21-b4c9-172a501374fd","Type":"ContainerStarted","Data":"23fd59a041d2389e984fa86d7f7a4fbb0400326dc272c36ea232f35fe96b1111"} Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.427588 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerStarted","Data":"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd"} Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.464327 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f241-account-create-rk5l6"] Sep 29 10:42:14 crc kubenswrapper[4727]: I0929 10:42:14.773625 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.129639 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f00fce5b-a355-42e2-94c7-63810205536f" path="/var/lib/kubelet/pods/f00fce5b-a355-42e2-94c7-63810205536f/volumes" Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.446670 4727 generic.go:334] "Generic (PLEG): container finished" podID="63978328-20b2-46a4-a423-b9c15c0c03cd" containerID="fd0caed0286b93020f6a5945968a273ec5c0414a5b208da2d9f07a4d1ba2da6d" exitCode=0 Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.447036 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad4a-account-create-blgkd" event={"ID":"63978328-20b2-46a4-a423-b9c15c0c03cd","Type":"ContainerDied","Data":"fd0caed0286b93020f6a5945968a273ec5c0414a5b208da2d9f07a4d1ba2da6d"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.452785 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56bbd245-faf2-4c21-b4c9-172a501374fd","Type":"ContainerStarted","Data":"e1fb9ba0b7230a935edfb7c42ed149baf6030b49b80dea359de2a81980335556"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.458663 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerStarted","Data":"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.474222 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230","Type":"ContainerStarted","Data":"b7fe319cb55813f94b9c833a5cc0072a023512e96582ef21ba261b92805bb5cb"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.474279 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230","Type":"ContainerStarted","Data":"d592af981312ef3b98b86922e88c4d5309d4f179c4eaff00b66018b423a193b3"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.478036 4727 generic.go:334] "Generic (PLEG): container finished" podID="2eedcefb-dda0-4045-abbd-611299c470aa" containerID="098ce89618430a91cc49dd15aef7f0be37b7a36fd22c333252cc8208db10684c" exitCode=0 Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.478092 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f241-account-create-rk5l6" event={"ID":"2eedcefb-dda0-4045-abbd-611299c470aa","Type":"ContainerDied","Data":"098ce89618430a91cc49dd15aef7f0be37b7a36fd22c333252cc8208db10684c"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.478124 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f241-account-create-rk5l6" event={"ID":"2eedcefb-dda0-4045-abbd-611299c470aa","Type":"ContainerStarted","Data":"4914b615febd5a0b64b9c9a50f0f54b37409b0c2e5b23d19666d58c6fb51e980"} Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.605117 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 29 10:42:15 crc kubenswrapper[4727]: I0929 10:42:15.969983 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.045714 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68ff49cc97-zxm88" Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.132944 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.133937 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-666c5cdbc4-zl8tq" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-httpd" containerID="cri-o://ae72109ce362862b0c543a6a98284e2ba0c67c420ef27519665a0e4e0fd27d94" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.133603 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-666c5cdbc4-zl8tq" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-api" containerID="cri-o://4845921ebcc5c86479aa47e1c246cdc0284636f6e5280acad46fd47666235adf" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.488854 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b1ab9d58-dea7-47fc-a259-5b9b8f7a7230","Type":"ContainerStarted","Data":"ff8f8ab43155d8e2575fc1e07c51234e45e986691e324aa8f5794fceea5d51ea"} Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.504293 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56bbd245-faf2-4c21-b4c9-172a501374fd","Type":"ContainerStarted","Data":"9a9c19a65c31612eddfacb79d7060f018138f48aa8b497b363e79d2a12aafef6"} Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507083 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-central-agent" containerID="cri-o://2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507312 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerStarted","Data":"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132"} Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507437 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507481 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="proxy-httpd" containerID="cri-o://6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507516 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="sg-core" containerID="cri-o://5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.507564 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-notification-agent" containerID="cri-o://867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd" gracePeriod=30 Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.529111 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.5290891269999998 podStartE2EDuration="3.529089127s" podCreationTimestamp="2025-09-29 10:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:42:16.52024681 +0000 UTC m=+1206.693560182" watchObservedRunningTime="2025-09-29 10:42:16.529089127 +0000 UTC m=+1206.702402489" Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.547640 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.57246427 podStartE2EDuration="6.547626034s" podCreationTimestamp="2025-09-29 10:42:10 +0000 UTC" firstStartedPulling="2025-09-29 10:42:11.193456309 +0000 UTC m=+1201.366769661" lastFinishedPulling="2025-09-29 10:42:16.168618063 +0000 UTC m=+1206.341931425" observedRunningTime="2025-09-29 10:42:16.544055792 +0000 UTC m=+1206.717369154" watchObservedRunningTime="2025-09-29 10:42:16.547626034 +0000 UTC m=+1206.720939386" Sep 29 10:42:16 crc kubenswrapper[4727]: I0929 10:42:16.643385 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.643360707 podStartE2EDuration="4.643360707s" podCreationTimestamp="2025-09-29 10:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:42:16.598945385 +0000 UTC m=+1206.772258747" watchObservedRunningTime="2025-09-29 10:42:16.643360707 +0000 UTC m=+1206.816674069" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.090331 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.114262 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.133156 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xjw9\" (UniqueName: \"kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9\") pod \"2eedcefb-dda0-4045-abbd-611299c470aa\" (UID: \"2eedcefb-dda0-4045-abbd-611299c470aa\") " Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.155580 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9" (OuterVolumeSpecName: "kube-api-access-7xjw9") pod "2eedcefb-dda0-4045-abbd-611299c470aa" (UID: "2eedcefb-dda0-4045-abbd-611299c470aa"). InnerVolumeSpecName "kube-api-access-7xjw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.234916 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k79ff\" (UniqueName: \"kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff\") pod \"63978328-20b2-46a4-a423-b9c15c0c03cd\" (UID: \"63978328-20b2-46a4-a423-b9c15c0c03cd\") " Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.235529 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xjw9\" (UniqueName: \"kubernetes.io/projected/2eedcefb-dda0-4045-abbd-611299c470aa-kube-api-access-7xjw9\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.240565 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff" (OuterVolumeSpecName: "kube-api-access-k79ff") pod "63978328-20b2-46a4-a423-b9c15c0c03cd" (UID: "63978328-20b2-46a4-a423-b9c15c0c03cd"). InnerVolumeSpecName "kube-api-access-k79ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.337610 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k79ff\" (UniqueName: \"kubernetes.io/projected/63978328-20b2-46a4-a423-b9c15c0c03cd-kube-api-access-k79ff\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.518723 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad4a-account-create-blgkd" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.519096 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad4a-account-create-blgkd" event={"ID":"63978328-20b2-46a4-a423-b9c15c0c03cd","Type":"ContainerDied","Data":"176d60411f9e00549cb3f23d70594fa3dc936bc83ff06c1e623a9854a29c429d"} Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.519147 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="176d60411f9e00549cb3f23d70594fa3dc936bc83ff06c1e623a9854a29c429d" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.521005 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f241-account-create-rk5l6" event={"ID":"2eedcefb-dda0-4045-abbd-611299c470aa","Type":"ContainerDied","Data":"4914b615febd5a0b64b9c9a50f0f54b37409b0c2e5b23d19666d58c6fb51e980"} Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.521035 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4914b615febd5a0b64b9c9a50f0f54b37409b0c2e5b23d19666d58c6fb51e980" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.521010 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f241-account-create-rk5l6" Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.525717 4727 generic.go:334] "Generic (PLEG): container finished" podID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerID="ae72109ce362862b0c543a6a98284e2ba0c67c420ef27519665a0e4e0fd27d94" exitCode=0 Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.525781 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerDied","Data":"ae72109ce362862b0c543a6a98284e2ba0c67c420ef27519665a0e4e0fd27d94"} Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528402 4727 generic.go:334] "Generic (PLEG): container finished" podID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerID="6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132" exitCode=0 Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528425 4727 generic.go:334] "Generic (PLEG): container finished" podID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerID="5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14" exitCode=2 Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528432 4727 generic.go:334] "Generic (PLEG): container finished" podID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerID="867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd" exitCode=0 Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528441 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerDied","Data":"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132"} Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528472 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerDied","Data":"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14"} Sep 29 10:42:17 crc kubenswrapper[4727]: I0929 10:42:17.528487 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerDied","Data":"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd"} Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.459259 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d7ljp"] Sep 29 10:42:18 crc kubenswrapper[4727]: E0929 10:42:18.459812 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63978328-20b2-46a4-a423-b9c15c0c03cd" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.459838 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="63978328-20b2-46a4-a423-b9c15c0c03cd" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: E0929 10:42:18.459859 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eedcefb-dda0-4045-abbd-611299c470aa" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.459867 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eedcefb-dda0-4045-abbd-611299c470aa" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.460094 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="63978328-20b2-46a4-a423-b9c15c0c03cd" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.460132 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eedcefb-dda0-4045-abbd-611299c470aa" containerName="mariadb-account-create" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.463024 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.465473 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.465607 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zg5gs" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.466150 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.467836 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d7ljp"] Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.565401 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.565452 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.565502 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6pwm\" (UniqueName: \"kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.565545 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.667388 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.667438 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.667497 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6pwm\" (UniqueName: \"kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.667549 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.676027 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.689868 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6pwm\" (UniqueName: \"kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.689924 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.690286 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data\") pod \"nova-cell0-conductor-db-sync-d7ljp\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:18 crc kubenswrapper[4727]: I0929 10:42:18.793043 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.321707 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d7ljp"] Sep 29 10:42:19 crc kubenswrapper[4727]: W0929 10:42:19.325660 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3866d914_1e2b_43e2_bf65_93d5e506fae2.slice/crio-693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa WatchSource:0}: Error finding container 693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa: Status 404 returned error can't find the container with id 693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.516186 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.557370 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" event={"ID":"3866d914-1e2b-43e2-bf65-93d5e506fae2","Type":"ContainerStarted","Data":"693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa"} Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.564953 4727 generic.go:334] "Generic (PLEG): container finished" podID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerID="2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da" exitCode=0 Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.565000 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerDied","Data":"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da"} Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.565028 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"059afe3c-c6ef-4e10-a186-f4407bc9ae41","Type":"ContainerDied","Data":"7f75a3d2efb9a6d31cfab993e62838f21bcc2577c88dbff80a08db86e1d5c620"} Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.565049 4727 scope.go:117] "RemoveContainer" containerID="6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.565198 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.587503 4727 scope.go:117] "RemoveContainer" containerID="5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597386 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597509 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597591 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597673 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597700 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv2kc\" (UniqueName: \"kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597753 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.597824 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml\") pod \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\" (UID: \"059afe3c-c6ef-4e10-a186-f4407bc9ae41\") " Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.604067 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.605153 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.611197 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts" (OuterVolumeSpecName: "scripts") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.611272 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc" (OuterVolumeSpecName: "kube-api-access-wv2kc") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "kube-api-access-wv2kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.626140 4727 scope.go:117] "RemoveContainer" containerID="867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.642638 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.653466 4727 scope.go:117] "RemoveContainer" containerID="2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.679892 4727 scope.go:117] "RemoveContainer" containerID="6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.680700 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132\": container with ID starting with 6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132 not found: ID does not exist" containerID="6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.680749 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132"} err="failed to get container status \"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132\": rpc error: code = NotFound desc = could not find container \"6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132\": container with ID starting with 6423caa8d0484479c2ca46f194a747fbe3cd8b812b210288960ba27ebed4c132 not found: ID does not exist" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.680779 4727 scope.go:117] "RemoveContainer" containerID="5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.683893 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14\": container with ID starting with 5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14 not found: ID does not exist" containerID="5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.683944 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14"} err="failed to get container status \"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14\": rpc error: code = NotFound desc = could not find container \"5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14\": container with ID starting with 5571951c2167ffcfb567b051b437994684dfe92d17ce20d9cb21b11bce2edb14 not found: ID does not exist" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.683980 4727 scope.go:117] "RemoveContainer" containerID="867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.685756 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd\": container with ID starting with 867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd not found: ID does not exist" containerID="867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.685823 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd"} err="failed to get container status \"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd\": rpc error: code = NotFound desc = could not find container \"867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd\": container with ID starting with 867dcf4004b18b9cd430126f588f586ecbfb42eb8ce385c925a8c93ed83755bd not found: ID does not exist" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.685858 4727 scope.go:117] "RemoveContainer" containerID="2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.686174 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.686221 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da\": container with ID starting with 2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da not found: ID does not exist" containerID="2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.686259 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da"} err="failed to get container status \"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da\": rpc error: code = NotFound desc = could not find container \"2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da\": container with ID starting with 2f6fbdcf145798f3994932b2d2806a85080e547cbfd6db96a10b6f7bcbe482da not found: ID does not exist" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.699976 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.700010 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.700021 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.700031 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.700042 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/059afe3c-c6ef-4e10-a186-f4407bc9ae41-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.700051 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv2kc\" (UniqueName: \"kubernetes.io/projected/059afe3c-c6ef-4e10-a186-f4407bc9ae41-kube-api-access-wv2kc\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.710544 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data" (OuterVolumeSpecName: "config-data") pod "059afe3c-c6ef-4e10-a186-f4407bc9ae41" (UID: "059afe3c-c6ef-4e10-a186-f4407bc9ae41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.801982 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059afe3c-c6ef-4e10-a186-f4407bc9ae41-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.899709 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.909929 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923183 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.923590 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-central-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923607 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-central-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.923643 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="sg-core" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923650 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="sg-core" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.923657 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-notification-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923664 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-notification-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: E0929 10:42:19.923673 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="proxy-httpd" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923678 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="proxy-httpd" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923849 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-central-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923860 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="sg-core" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923884 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="proxy-httpd" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.923892 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" containerName="ceilometer-notification-agent" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.925559 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.930437 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.931127 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:42:19 crc kubenswrapper[4727]: I0929 10:42:19.932710 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.005559 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxwmt\" (UniqueName: \"kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.005614 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.005644 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.005669 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.005909 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.006056 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.006181 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108127 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108229 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108285 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108355 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108414 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxwmt\" (UniqueName: \"kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108458 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108490 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.108974 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.109024 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.113987 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.114071 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.118616 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.119095 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.129132 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxwmt\" (UniqueName: \"kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt\") pod \"ceilometer-0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.241487 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.738021 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:20 crc kubenswrapper[4727]: I0929 10:42:20.871023 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 29 10:42:21 crc kubenswrapper[4727]: I0929 10:42:21.122330 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="059afe3c-c6ef-4e10-a186-f4407bc9ae41" path="/var/lib/kubelet/pods/059afe3c-c6ef-4e10-a186-f4407bc9ae41/volumes" Sep 29 10:42:21 crc kubenswrapper[4727]: I0929 10:42:21.591485 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerStarted","Data":"67c7252cec796a7a20f060d7d4f80234872610bfca4c87c262152c5141845837"} Sep 29 10:42:21 crc kubenswrapper[4727]: I0929 10:42:21.591840 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerStarted","Data":"b0e5ce2ffc0437f5eb3df1cdc635fb8130c7970a3a1147356478dcb37cd6c4fa"} Sep 29 10:42:22 crc kubenswrapper[4727]: I0929 10:42:22.604296 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerStarted","Data":"e4a48047a2351b17bc778a297af84bb7aeb210f1138bee254c482168cceb4e76"} Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.121093 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.121136 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.160823 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.191621 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.627220 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.627617 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.849302 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.849361 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.898199 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:23 crc kubenswrapper[4727]: I0929 10:42:23.902632 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:24 crc kubenswrapper[4727]: I0929 10:42:24.637158 4727 generic.go:334] "Generic (PLEG): container finished" podID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerID="4845921ebcc5c86479aa47e1c246cdc0284636f6e5280acad46fd47666235adf" exitCode=0 Sep 29 10:42:24 crc kubenswrapper[4727]: I0929 10:42:24.638369 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerDied","Data":"4845921ebcc5c86479aa47e1c246cdc0284636f6e5280acad46fd47666235adf"} Sep 29 10:42:24 crc kubenswrapper[4727]: I0929 10:42:24.638429 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:24 crc kubenswrapper[4727]: I0929 10:42:24.638899 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:25 crc kubenswrapper[4727]: I0929 10:42:25.966632 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 10:42:25 crc kubenswrapper[4727]: I0929 10:42:25.967081 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:42:25 crc kubenswrapper[4727]: I0929 10:42:25.967902 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 29 10:42:26 crc kubenswrapper[4727]: I0929 10:42:26.642890 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:26 crc kubenswrapper[4727]: I0929 10:42:26.656834 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:42:26 crc kubenswrapper[4727]: I0929 10:42:26.656862 4727 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 10:42:26 crc kubenswrapper[4727]: I0929 10:42:26.854043 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:26 crc kubenswrapper[4727]: I0929 10:42:26.860566 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.652706 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.683993 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666c5cdbc4-zl8tq" event={"ID":"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a","Type":"ContainerDied","Data":"506e7dbef15528cca921ab2e9016a81743c2931693174f0ae3b56cdb05835c27"} Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.684035 4727 scope.go:117] "RemoveContainer" containerID="ae72109ce362862b0c543a6a98284e2ba0c67c420ef27519665a0e4e0fd27d94" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.684130 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666c5cdbc4-zl8tq" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.686055 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" event={"ID":"3866d914-1e2b-43e2-bf65-93d5e506fae2","Type":"ContainerStarted","Data":"f6adc8347eccca98ece20adbd665742a43a470278d1c0f9dcb1f598e250ae4a9"} Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.688601 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerStarted","Data":"5272d13d7f54b139c514b988b5034dbab548028f5bf1270343b36770cb8d2121"} Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.704667 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" podStartSLOduration=1.7036572840000002 podStartE2EDuration="11.704649197s" podCreationTimestamp="2025-09-29 10:42:18 +0000 UTC" firstStartedPulling="2025-09-29 10:42:19.327964348 +0000 UTC m=+1209.501277710" lastFinishedPulling="2025-09-29 10:42:29.328956261 +0000 UTC m=+1219.502269623" observedRunningTime="2025-09-29 10:42:29.703744043 +0000 UTC m=+1219.877057405" watchObservedRunningTime="2025-09-29 10:42:29.704649197 +0000 UTC m=+1219.877962559" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.709945 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle\") pod \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.710021 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhwcf\" (UniqueName: \"kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf\") pod \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.717307 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf" (OuterVolumeSpecName: "kube-api-access-qhwcf") pod "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" (UID: "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a"). InnerVolumeSpecName "kube-api-access-qhwcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.732547 4727 scope.go:117] "RemoveContainer" containerID="4845921ebcc5c86479aa47e1c246cdc0284636f6e5280acad46fd47666235adf" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.777205 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" (UID: "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.812238 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config\") pod \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.813027 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs\") pod \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.813200 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config\") pod \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\" (UID: \"8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a\") " Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.814011 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.814131 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhwcf\" (UniqueName: \"kubernetes.io/projected/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-kube-api-access-qhwcf\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.815924 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" (UID: "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.873466 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config" (OuterVolumeSpecName: "config") pod "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" (UID: "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.881099 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" (UID: "8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.914872 4727 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.914907 4727 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:29 crc kubenswrapper[4727]: I0929 10:42:29.914918 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:30 crc kubenswrapper[4727]: I0929 10:42:30.020117 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:42:30 crc kubenswrapper[4727]: I0929 10:42:30.030612 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-666c5cdbc4-zl8tq"] Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.135582 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" path="/var/lib/kubelet/pods/8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a/volumes" Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.718584 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerStarted","Data":"ebf50f95321bb3eb023e0906cd29514e263856656a723819d07977c5df164b1e"} Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.719017 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-central-agent" containerID="cri-o://67c7252cec796a7a20f060d7d4f80234872610bfca4c87c262152c5141845837" gracePeriod=30 Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.719434 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.719685 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="proxy-httpd" containerID="cri-o://ebf50f95321bb3eb023e0906cd29514e263856656a723819d07977c5df164b1e" gracePeriod=30 Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.719729 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="sg-core" containerID="cri-o://5272d13d7f54b139c514b988b5034dbab548028f5bf1270343b36770cb8d2121" gracePeriod=30 Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.719778 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-notification-agent" containerID="cri-o://e4a48047a2351b17bc778a297af84bb7aeb210f1138bee254c482168cceb4e76" gracePeriod=30 Sep 29 10:42:31 crc kubenswrapper[4727]: I0929 10:42:31.741431 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.336609479 podStartE2EDuration="12.740305612s" podCreationTimestamp="2025-09-29 10:42:19 +0000 UTC" firstStartedPulling="2025-09-29 10:42:20.760500686 +0000 UTC m=+1210.933814058" lastFinishedPulling="2025-09-29 10:42:31.164196829 +0000 UTC m=+1221.337510191" observedRunningTime="2025-09-29 10:42:31.737402297 +0000 UTC m=+1221.910715679" watchObservedRunningTime="2025-09-29 10:42:31.740305612 +0000 UTC m=+1221.913618984" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.730933 4727 generic.go:334] "Generic (PLEG): container finished" podID="83e01431-9844-4359-a27a-739f29333ad0" containerID="ebf50f95321bb3eb023e0906cd29514e263856656a723819d07977c5df164b1e" exitCode=0 Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731314 4727 generic.go:334] "Generic (PLEG): container finished" podID="83e01431-9844-4359-a27a-739f29333ad0" containerID="5272d13d7f54b139c514b988b5034dbab548028f5bf1270343b36770cb8d2121" exitCode=2 Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731325 4727 generic.go:334] "Generic (PLEG): container finished" podID="83e01431-9844-4359-a27a-739f29333ad0" containerID="e4a48047a2351b17bc778a297af84bb7aeb210f1138bee254c482168cceb4e76" exitCode=0 Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731347 4727 generic.go:334] "Generic (PLEG): container finished" podID="83e01431-9844-4359-a27a-739f29333ad0" containerID="67c7252cec796a7a20f060d7d4f80234872610bfca4c87c262152c5141845837" exitCode=0 Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731010 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerDied","Data":"ebf50f95321bb3eb023e0906cd29514e263856656a723819d07977c5df164b1e"} Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731393 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerDied","Data":"5272d13d7f54b139c514b988b5034dbab548028f5bf1270343b36770cb8d2121"} Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731411 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerDied","Data":"e4a48047a2351b17bc778a297af84bb7aeb210f1138bee254c482168cceb4e76"} Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731424 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerDied","Data":"67c7252cec796a7a20f060d7d4f80234872610bfca4c87c262152c5141845837"} Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731435 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83e01431-9844-4359-a27a-739f29333ad0","Type":"ContainerDied","Data":"b0e5ce2ffc0437f5eb3df1cdc635fb8130c7970a3a1147356478dcb37cd6c4fa"} Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.731446 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0e5ce2ffc0437f5eb3df1cdc635fb8130c7970a3a1147356478dcb37cd6c4fa" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.783801 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986079 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986511 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986528 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986558 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986679 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986740 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxwmt\" (UniqueName: \"kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.986786 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle\") pod \"83e01431-9844-4359-a27a-739f29333ad0\" (UID: \"83e01431-9844-4359-a27a-739f29333ad0\") " Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.987150 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.987173 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.991345 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts" (OuterVolumeSpecName: "scripts") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:32 crc kubenswrapper[4727]: I0929 10:42:32.992452 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt" (OuterVolumeSpecName: "kube-api-access-nxwmt") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "kube-api-access-nxwmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.025097 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.062641 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.088943 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.089169 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.089259 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83e01431-9844-4359-a27a-739f29333ad0-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.089321 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.089420 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxwmt\" (UniqueName: \"kubernetes.io/projected/83e01431-9844-4359-a27a-739f29333ad0-kube-api-access-nxwmt\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.089481 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.099140 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data" (OuterVolumeSpecName: "config-data") pod "83e01431-9844-4359-a27a-739f29333ad0" (UID: "83e01431-9844-4359-a27a-739f29333ad0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.191242 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e01431-9844-4359-a27a-739f29333ad0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.742044 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.781378 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.790507 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.798470 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.799090 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-notification-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.799167 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-notification-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.799240 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-central-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.799291 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-central-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.799374 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800006 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.800080 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="sg-core" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800143 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="sg-core" Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.800208 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-api" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800258 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-api" Sep 29 10:42:33 crc kubenswrapper[4727]: E0929 10:42:33.800309 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="proxy-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800377 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="proxy-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800732 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.800812 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-notification-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.801606 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="sg-core" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.801652 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="proxy-httpd" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.801670 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f6bd456-e04d-4dcc-9c57-d3f8e7c9919a" containerName="neutron-api" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.801687 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e01431-9844-4359-a27a-739f29333ad0" containerName="ceilometer-central-agent" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.803919 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.808924 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.847210 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.847659 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.905917 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906065 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906121 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906201 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906223 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906263 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:33 crc kubenswrapper[4727]: I0929 10:42:33.906280 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j9ks\" (UniqueName: \"kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007549 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007650 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007675 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007725 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007744 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007768 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.007783 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j9ks\" (UniqueName: \"kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.008970 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.009168 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.013578 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.014388 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.015211 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.017226 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.027918 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j9ks\" (UniqueName: \"kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks\") pod \"ceilometer-0\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.168264 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.470660 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.638243 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:34 crc kubenswrapper[4727]: I0929 10:42:34.754908 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerStarted","Data":"cd05fce053a41a192f9562ac5af6c6d78e4f59515e93950a41fe98d29a2e52c3"} Sep 29 10:42:35 crc kubenswrapper[4727]: I0929 10:42:35.120159 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e01431-9844-4359-a27a-739f29333ad0" path="/var/lib/kubelet/pods/83e01431-9844-4359-a27a-739f29333ad0/volumes" Sep 29 10:42:37 crc kubenswrapper[4727]: I0929 10:42:37.784491 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerStarted","Data":"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224"} Sep 29 10:42:37 crc kubenswrapper[4727]: I0929 10:42:37.785065 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerStarted","Data":"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1"} Sep 29 10:42:38 crc kubenswrapper[4727]: I0929 10:42:38.796615 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerStarted","Data":"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01"} Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.840538 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerStarted","Data":"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01"} Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.841101 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.840891 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="proxy-httpd" containerID="cri-o://46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01" gracePeriod=30 Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.840670 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-central-agent" containerID="cri-o://d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1" gracePeriod=30 Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.840945 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="sg-core" containerID="cri-o://616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01" gracePeriod=30 Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.840933 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-notification-agent" containerID="cri-o://ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224" gracePeriod=30 Sep 29 10:42:43 crc kubenswrapper[4727]: I0929 10:42:43.866247 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.387738735 podStartE2EDuration="10.866231484s" podCreationTimestamp="2025-09-29 10:42:33 +0000 UTC" firstStartedPulling="2025-09-29 10:42:34.650268651 +0000 UTC m=+1224.823582013" lastFinishedPulling="2025-09-29 10:42:43.12876139 +0000 UTC m=+1233.302074762" observedRunningTime="2025-09-29 10:42:43.863533064 +0000 UTC m=+1234.036846426" watchObservedRunningTime="2025-09-29 10:42:43.866231484 +0000 UTC m=+1234.039544846" Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851102 4727 generic.go:334] "Generic (PLEG): container finished" podID="841030a1-cd45-44d2-921c-ba5beb088aff" containerID="46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01" exitCode=0 Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851483 4727 generic.go:334] "Generic (PLEG): container finished" podID="841030a1-cd45-44d2-921c-ba5beb088aff" containerID="616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01" exitCode=2 Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851496 4727 generic.go:334] "Generic (PLEG): container finished" podID="841030a1-cd45-44d2-921c-ba5beb088aff" containerID="ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224" exitCode=0 Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851170 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerDied","Data":"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01"} Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851534 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerDied","Data":"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01"} Sep 29 10:42:44 crc kubenswrapper[4727]: I0929 10:42:44.851552 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerDied","Data":"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224"} Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.688481 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.845856 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j9ks\" (UniqueName: \"kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.845904 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.845949 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.846061 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.846089 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.846131 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.846178 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd\") pod \"841030a1-cd45-44d2-921c-ba5beb088aff\" (UID: \"841030a1-cd45-44d2-921c-ba5beb088aff\") " Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.846827 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.847266 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.852646 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks" (OuterVolumeSpecName: "kube-api-access-9j9ks") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "kube-api-access-9j9ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.860436 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts" (OuterVolumeSpecName: "scripts") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.872922 4727 generic.go:334] "Generic (PLEG): container finished" podID="841030a1-cd45-44d2-921c-ba5beb088aff" containerID="d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1" exitCode=0 Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.872968 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerDied","Data":"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1"} Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.872995 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"841030a1-cd45-44d2-921c-ba5beb088aff","Type":"ContainerDied","Data":"cd05fce053a41a192f9562ac5af6c6d78e4f59515e93950a41fe98d29a2e52c3"} Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.873013 4727 scope.go:117] "RemoveContainer" containerID="46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.873144 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.877444 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.933362 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.946626 4727 scope.go:117] "RemoveContainer" containerID="616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947697 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j9ks\" (UniqueName: \"kubernetes.io/projected/841030a1-cd45-44d2-921c-ba5beb088aff-kube-api-access-9j9ks\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947731 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947740 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947748 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947761 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.947770 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/841030a1-cd45-44d2-921c-ba5beb088aff-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.950834 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data" (OuterVolumeSpecName: "config-data") pod "841030a1-cd45-44d2-921c-ba5beb088aff" (UID: "841030a1-cd45-44d2-921c-ba5beb088aff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:42:45 crc kubenswrapper[4727]: I0929 10:42:45.964660 4727 scope.go:117] "RemoveContainer" containerID="ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.006551 4727 scope.go:117] "RemoveContainer" containerID="d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.033109 4727 scope.go:117] "RemoveContainer" containerID="46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.033596 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01\": container with ID starting with 46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01 not found: ID does not exist" containerID="46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.033637 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01"} err="failed to get container status \"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01\": rpc error: code = NotFound desc = could not find container \"46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01\": container with ID starting with 46cfc77fec99770d4cf5b74ae7a0812c30885b21aada68a1c62ce349fa476d01 not found: ID does not exist" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.033664 4727 scope.go:117] "RemoveContainer" containerID="616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.033984 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01\": container with ID starting with 616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01 not found: ID does not exist" containerID="616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.034020 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01"} err="failed to get container status \"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01\": rpc error: code = NotFound desc = could not find container \"616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01\": container with ID starting with 616e7d3db45c1a5f3ff6297f017ea0177d537c2a3889c2564fcdfa4580499b01 not found: ID does not exist" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.034042 4727 scope.go:117] "RemoveContainer" containerID="ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.034393 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224\": container with ID starting with ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224 not found: ID does not exist" containerID="ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.034434 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224"} err="failed to get container status \"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224\": rpc error: code = NotFound desc = could not find container \"ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224\": container with ID starting with ff3b4f8dd07b7cda1827e07e6990f48556080acb0eb7b5b3ccff49a175c46224 not found: ID does not exist" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.034448 4727 scope.go:117] "RemoveContainer" containerID="d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.034773 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1\": container with ID starting with d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1 not found: ID does not exist" containerID="d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.034803 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1"} err="failed to get container status \"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1\": rpc error: code = NotFound desc = could not find container \"d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1\": container with ID starting with d4da86a3df45d17d6a4bb92358325fb259588ccc5a5fbe5687b4f2e0682b1db1 not found: ID does not exist" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.049286 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841030a1-cd45-44d2-921c-ba5beb088aff-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.208669 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.220394 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.241511 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.242330 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="proxy-httpd" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.242360 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="proxy-httpd" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.242378 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-notification-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.242389 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-notification-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.242396 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="sg-core" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.242402 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="sg-core" Sep 29 10:42:46 crc kubenswrapper[4727]: E0929 10:42:46.242437 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-central-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.242443 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-central-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.244738 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="proxy-httpd" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.244773 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-notification-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.244794 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="sg-core" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.244801 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" containerName="ceilometer-central-agent" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.267681 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.267796 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.272134 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.272477 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460516 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460585 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460658 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460702 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88bdz\" (UniqueName: \"kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460723 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460774 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.460804 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562583 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562639 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88bdz\" (UniqueName: \"kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562662 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562742 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562787 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562835 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.562868 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.563408 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.563738 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.568002 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.568177 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.568385 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.568546 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.579621 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88bdz\" (UniqueName: \"kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz\") pod \"ceilometer-0\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " pod="openstack/ceilometer-0" Sep 29 10:42:46 crc kubenswrapper[4727]: I0929 10:42:46.607003 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:42:47 crc kubenswrapper[4727]: I0929 10:42:47.049533 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:42:47 crc kubenswrapper[4727]: W0929 10:42:47.054152 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac6ad249_abc1_4de5_8210_ba66157bc5f6.slice/crio-79d9beb297e188f258179b4dfd2d1af80fd0b11c4de216efd4aa1baccdf0f025 WatchSource:0}: Error finding container 79d9beb297e188f258179b4dfd2d1af80fd0b11c4de216efd4aa1baccdf0f025: Status 404 returned error can't find the container with id 79d9beb297e188f258179b4dfd2d1af80fd0b11c4de216efd4aa1baccdf0f025 Sep 29 10:42:47 crc kubenswrapper[4727]: I0929 10:42:47.120260 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="841030a1-cd45-44d2-921c-ba5beb088aff" path="/var/lib/kubelet/pods/841030a1-cd45-44d2-921c-ba5beb088aff/volumes" Sep 29 10:42:47 crc kubenswrapper[4727]: I0929 10:42:47.895474 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerStarted","Data":"79d9beb297e188f258179b4dfd2d1af80fd0b11c4de216efd4aa1baccdf0f025"} Sep 29 10:42:48 crc kubenswrapper[4727]: I0929 10:42:48.912129 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerStarted","Data":"48df589cd381bb4c8bb998173003118441f21802c8afc32012966fbdc41baa78"} Sep 29 10:42:50 crc kubenswrapper[4727]: I0929 10:42:50.928187 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerStarted","Data":"71f325f1585194e04b800a7cc6972876d6bbea088a890589cc7ea0d55ef24492"} Sep 29 10:42:52 crc kubenswrapper[4727]: I0929 10:42:52.954638 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerStarted","Data":"d16983737a991b453b1cf8b8b0eaa67b15c90872c4017ec0f45b01377d0e9c53"} Sep 29 10:42:54 crc kubenswrapper[4727]: I0929 10:42:54.979929 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerStarted","Data":"3891805d5fdd737649f8ea893ba2b72448437cc39f2b0ae65726c379d8988405"} Sep 29 10:42:54 crc kubenswrapper[4727]: I0929 10:42:54.980708 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:42:55 crc kubenswrapper[4727]: I0929 10:42:55.004522 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.116658749 podStartE2EDuration="9.004502135s" podCreationTimestamp="2025-09-29 10:42:46 +0000 UTC" firstStartedPulling="2025-09-29 10:42:47.058151919 +0000 UTC m=+1237.231465281" lastFinishedPulling="2025-09-29 10:42:53.945995305 +0000 UTC m=+1244.119308667" observedRunningTime="2025-09-29 10:42:54.997932276 +0000 UTC m=+1245.171245638" watchObservedRunningTime="2025-09-29 10:42:55.004502135 +0000 UTC m=+1245.177815507" Sep 29 10:43:01 crc kubenswrapper[4727]: I0929 10:43:01.046924 4727 generic.go:334] "Generic (PLEG): container finished" podID="3866d914-1e2b-43e2-bf65-93d5e506fae2" containerID="f6adc8347eccca98ece20adbd665742a43a470278d1c0f9dcb1f598e250ae4a9" exitCode=0 Sep 29 10:43:01 crc kubenswrapper[4727]: I0929 10:43:01.047000 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" event={"ID":"3866d914-1e2b-43e2-bf65-93d5e506fae2","Type":"ContainerDied","Data":"f6adc8347eccca98ece20adbd665742a43a470278d1c0f9dcb1f598e250ae4a9"} Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.378211 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.456378 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6pwm\" (UniqueName: \"kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm\") pod \"3866d914-1e2b-43e2-bf65-93d5e506fae2\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.456439 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts\") pod \"3866d914-1e2b-43e2-bf65-93d5e506fae2\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.456458 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle\") pod \"3866d914-1e2b-43e2-bf65-93d5e506fae2\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.456543 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data\") pod \"3866d914-1e2b-43e2-bf65-93d5e506fae2\" (UID: \"3866d914-1e2b-43e2-bf65-93d5e506fae2\") " Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.461969 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm" (OuterVolumeSpecName: "kube-api-access-l6pwm") pod "3866d914-1e2b-43e2-bf65-93d5e506fae2" (UID: "3866d914-1e2b-43e2-bf65-93d5e506fae2"). InnerVolumeSpecName "kube-api-access-l6pwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.462486 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts" (OuterVolumeSpecName: "scripts") pod "3866d914-1e2b-43e2-bf65-93d5e506fae2" (UID: "3866d914-1e2b-43e2-bf65-93d5e506fae2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.485593 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data" (OuterVolumeSpecName: "config-data") pod "3866d914-1e2b-43e2-bf65-93d5e506fae2" (UID: "3866d914-1e2b-43e2-bf65-93d5e506fae2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.485921 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3866d914-1e2b-43e2-bf65-93d5e506fae2" (UID: "3866d914-1e2b-43e2-bf65-93d5e506fae2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.558898 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6pwm\" (UniqueName: \"kubernetes.io/projected/3866d914-1e2b-43e2-bf65-93d5e506fae2-kube-api-access-l6pwm\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.558929 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.558941 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:02 crc kubenswrapper[4727]: I0929 10:43:02.558951 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3866d914-1e2b-43e2-bf65-93d5e506fae2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.063494 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" event={"ID":"3866d914-1e2b-43e2-bf65-93d5e506fae2","Type":"ContainerDied","Data":"693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa"} Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.063546 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="693906443db710bef01342c5aef8032d39f0c9392988c69c5b64b962e013a2fa" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.063556 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d7ljp" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.179407 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 10:43:03 crc kubenswrapper[4727]: E0929 10:43:03.180172 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3866d914-1e2b-43e2-bf65-93d5e506fae2" containerName="nova-cell0-conductor-db-sync" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.180191 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="3866d914-1e2b-43e2-bf65-93d5e506fae2" containerName="nova-cell0-conductor-db-sync" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.180407 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="3866d914-1e2b-43e2-bf65-93d5e506fae2" containerName="nova-cell0-conductor-db-sync" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.181078 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.190091 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zg5gs" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.190434 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.204224 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.273756 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw4t5\" (UniqueName: \"kubernetes.io/projected/45ccefb3-e0b3-4484-af4b-29b81d687507-kube-api-access-jw4t5\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.275070 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.275106 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.376856 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw4t5\" (UniqueName: \"kubernetes.io/projected/45ccefb3-e0b3-4484-af4b-29b81d687507-kube-api-access-jw4t5\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.376987 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.377020 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.381531 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.381532 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ccefb3-e0b3-4484-af4b-29b81d687507-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.397667 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw4t5\" (UniqueName: \"kubernetes.io/projected/45ccefb3-e0b3-4484-af4b-29b81d687507-kube-api-access-jw4t5\") pod \"nova-cell0-conductor-0\" (UID: \"45ccefb3-e0b3-4484-af4b-29b81d687507\") " pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:03 crc kubenswrapper[4727]: I0929 10:43:03.558208 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:04 crc kubenswrapper[4727]: I0929 10:43:04.044598 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 29 10:43:04 crc kubenswrapper[4727]: W0929 10:43:04.051377 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45ccefb3_e0b3_4484_af4b_29b81d687507.slice/crio-363c6712bf098f1c1ae01ce7d93d3ee545680a57272b9b5d1c80802eba82641d WatchSource:0}: Error finding container 363c6712bf098f1c1ae01ce7d93d3ee545680a57272b9b5d1c80802eba82641d: Status 404 returned error can't find the container with id 363c6712bf098f1c1ae01ce7d93d3ee545680a57272b9b5d1c80802eba82641d Sep 29 10:43:04 crc kubenswrapper[4727]: I0929 10:43:04.072707 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"45ccefb3-e0b3-4484-af4b-29b81d687507","Type":"ContainerStarted","Data":"363c6712bf098f1c1ae01ce7d93d3ee545680a57272b9b5d1c80802eba82641d"} Sep 29 10:43:05 crc kubenswrapper[4727]: I0929 10:43:05.086961 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"45ccefb3-e0b3-4484-af4b-29b81d687507","Type":"ContainerStarted","Data":"4c7fcbe73c5ccb027449a8cbb1157c34ec720bbcc2da0a7c65a7cd729e55e445"} Sep 29 10:43:05 crc kubenswrapper[4727]: I0929 10:43:05.087431 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:05 crc kubenswrapper[4727]: I0929 10:43:05.125581 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.125559582 podStartE2EDuration="2.125559582s" podCreationTimestamp="2025-09-29 10:43:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:05.106675565 +0000 UTC m=+1255.279988927" watchObservedRunningTime="2025-09-29 10:43:05.125559582 +0000 UTC m=+1255.298872934" Sep 29 10:43:13 crc kubenswrapper[4727]: I0929 10:43:13.584156 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.166839 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-dl7pr"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.168578 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.171195 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.171848 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.185954 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dl7pr"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.295319 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.295523 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hdmr\" (UniqueName: \"kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.295740 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.295832 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.358068 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.359582 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.361919 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.371164 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.398468 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.398548 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.398609 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.398674 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hdmr\" (UniqueName: \"kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.413239 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.414022 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.415875 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.419178 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hdmr\" (UniqueName: \"kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr\") pod \"nova-cell0-cell-mapping-dl7pr\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.434458 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.435770 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.439629 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.440356 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.501694 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrhf9\" (UniqueName: \"kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.501768 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.501794 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.501873 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.501956 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.502028 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.502065 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hvtp\" (UniqueName: \"kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.529792 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.561577 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.563160 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.572924 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606642 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606720 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606748 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hvtp\" (UniqueName: \"kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606793 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrhf9\" (UniqueName: \"kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606812 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606832 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606861 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.606882 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.607462 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.615541 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.619033 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.628769 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.632378 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.661195 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.662612 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.670728 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hvtp\" (UniqueName: \"kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp\") pod \"nova-scheduler-0\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.671087 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.683906 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.691875 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrhf9\" (UniqueName: \"kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9\") pod \"nova-api-0\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " pod="openstack/nova-api-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.705482 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.706980 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.708289 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvs6c\" (UniqueName: \"kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.708357 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.708389 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.708472 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.732412 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810491 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810571 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810633 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfbgb\" (UniqueName: \"kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810663 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810904 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.810974 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811008 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811032 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvs6c\" (UniqueName: \"kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811055 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmhzc\" (UniqueName: \"kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811202 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811324 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811426 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811469 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.811948 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.815600 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.815763 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.832627 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.836008 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvs6c\" (UniqueName: \"kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c\") pod \"nova-metadata-0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.894547 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.916791 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.916867 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.916896 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfbgb\" (UniqueName: \"kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.916945 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.917095 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.917124 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.917153 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.917179 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmhzc\" (UniqueName: \"kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.917209 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.918624 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.919520 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.921520 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.921715 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.924016 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.924459 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.926578 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.934325 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfbgb\" (UniqueName: \"kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.937432 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmhzc\" (UniqueName: \"kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc\") pod \"dnsmasq-dns-699d7ddff-6tqz6\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:14 crc kubenswrapper[4727]: I0929 10:43:14.979109 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.071775 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.083397 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.289915 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dl7pr"] Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.357423 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8n5tt"] Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.358745 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.364919 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.365794 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.373288 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8n5tt"] Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.428024 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.429282 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlt9c\" (UniqueName: \"kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.432566 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.432832 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.433028 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.472477 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.488189 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:15 crc kubenswrapper[4727]: W0929 10:43:15.519568 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13e791f8_a90a_4d87_934a_9bb569c9d6a7.slice/crio-edc4da5a6b93578415f2eacc2904a322e7e4ec3246d572a05764e88a6f358092 WatchSource:0}: Error finding container edc4da5a6b93578415f2eacc2904a322e7e4ec3246d572a05764e88a6f358092: Status 404 returned error can't find the container with id edc4da5a6b93578415f2eacc2904a322e7e4ec3246d572a05764e88a6f358092 Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.537811 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlt9c\" (UniqueName: \"kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.537861 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.537889 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.537950 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.554380 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.556136 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.569183 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlt9c\" (UniqueName: \"kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.569678 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data\") pod \"nova-cell1-conductor-db-sync-8n5tt\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.685765 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.843754 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:15 crc kubenswrapper[4727]: W0929 10:43:15.852294 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1232ec7_a034_4ed1_845b_0ffc5bb900d7.slice/crio-424d1fcd8323664b1ff76a9cf65a9ca920321c2b85245c80896d52cf3c9cb835 WatchSource:0}: Error finding container 424d1fcd8323664b1ff76a9cf65a9ca920321c2b85245c80896d52cf3c9cb835: Status 404 returned error can't find the container with id 424d1fcd8323664b1ff76a9cf65a9ca920321c2b85245c80896d52cf3c9cb835 Sep 29 10:43:15 crc kubenswrapper[4727]: I0929 10:43:15.880901 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:43:15 crc kubenswrapper[4727]: W0929 10:43:15.888877 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd841b62_c07f_47bc_8da0_1f36a1e51176.slice/crio-67774f38e7b0d0a8ae9fa72d82d3dd51a1502b17c8f363a2cda0d5c2a0f9cae3 WatchSource:0}: Error finding container 67774f38e7b0d0a8ae9fa72d82d3dd51a1502b17c8f363a2cda0d5c2a0f9cae3: Status 404 returned error can't find the container with id 67774f38e7b0d0a8ae9fa72d82d3dd51a1502b17c8f363a2cda0d5c2a0f9cae3 Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.184861 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0fca1e2c-1ea1-4239-aae1-6da063df9c3a","Type":"ContainerStarted","Data":"3f7e9683ee4f1afde98bbc60f4121b6a6e16fc33d69c0b340d881c8c50df69db"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.186937 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerStarted","Data":"407346dd0b039458a59e185d34fe7429e16bca7aac4b43154d9c47b35e321f67"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.186976 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerStarted","Data":"67774f38e7b0d0a8ae9fa72d82d3dd51a1502b17c8f363a2cda0d5c2a0f9cae3"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.191281 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1232ec7-a034-4ed1-845b-0ffc5bb900d7","Type":"ContainerStarted","Data":"424d1fcd8323664b1ff76a9cf65a9ca920321c2b85245c80896d52cf3c9cb835"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.193253 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerStarted","Data":"edc4da5a6b93578415f2eacc2904a322e7e4ec3246d572a05764e88a6f358092"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.194693 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerStarted","Data":"50ade81c394c316f0d0f5e6af22f0e540366547eac2033465a287746fa298648"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.202078 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dl7pr" event={"ID":"da99e317-fa23-4a56-a23d-1447b1ad3cb0","Type":"ContainerStarted","Data":"a34b5738bfed82f755ab116015123779cad6e88fe1eb7e97a94f7e491172bf58"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.202147 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dl7pr" event={"ID":"da99e317-fa23-4a56-a23d-1447b1ad3cb0","Type":"ContainerStarted","Data":"bf64ed6def6b24de87cb6b79886b86ea4be49e29c065c0e5ba7497b59f902ec1"} Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.224761 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-dl7pr" podStartSLOduration=2.224743635 podStartE2EDuration="2.224743635s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:16.223234656 +0000 UTC m=+1266.396548028" watchObservedRunningTime="2025-09-29 10:43:16.224743635 +0000 UTC m=+1266.398056997" Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.264940 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8n5tt"] Sep 29 10:43:16 crc kubenswrapper[4727]: W0929 10:43:16.265798 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6027aeb9_0794_453e_85e3_9bf0e0c35d61.slice/crio-cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c WatchSource:0}: Error finding container cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c: Status 404 returned error can't find the container with id cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c Sep 29 10:43:16 crc kubenswrapper[4727]: I0929 10:43:16.628661 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 10:43:17 crc kubenswrapper[4727]: I0929 10:43:17.219644 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" event={"ID":"6027aeb9-0794-453e-85e3-9bf0e0c35d61","Type":"ContainerStarted","Data":"4cd2c1e0dcfb72c0d3f22f6f8d299f86231c1885b308943dd7207aea3df69e55"} Sep 29 10:43:17 crc kubenswrapper[4727]: I0929 10:43:17.219692 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" event={"ID":"6027aeb9-0794-453e-85e3-9bf0e0c35d61","Type":"ContainerStarted","Data":"cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c"} Sep 29 10:43:17 crc kubenswrapper[4727]: I0929 10:43:17.221435 4727 generic.go:334] "Generic (PLEG): container finished" podID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerID="407346dd0b039458a59e185d34fe7429e16bca7aac4b43154d9c47b35e321f67" exitCode=0 Sep 29 10:43:17 crc kubenswrapper[4727]: I0929 10:43:17.225119 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerDied","Data":"407346dd0b039458a59e185d34fe7429e16bca7aac4b43154d9c47b35e321f67"} Sep 29 10:43:17 crc kubenswrapper[4727]: I0929 10:43:17.303671 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" podStartSLOduration=2.303650783 podStartE2EDuration="2.303650783s" podCreationTimestamp="2025-09-29 10:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:17.252639937 +0000 UTC m=+1267.425953299" watchObservedRunningTime="2025-09-29 10:43:17.303650783 +0000 UTC m=+1267.476964145" Sep 29 10:43:18 crc kubenswrapper[4727]: I0929 10:43:18.324205 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:18 crc kubenswrapper[4727]: I0929 10:43:18.380399 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.258799 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerStarted","Data":"aee4fd018b4ccd735a6280897bf356facb5eb8006522a7d66a8b7d462d741804"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.259431 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.261633 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1232ec7-a034-4ed1-845b-0ffc5bb900d7","Type":"ContainerStarted","Data":"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.261636 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349" gracePeriod=30 Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.265188 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerStarted","Data":"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.265239 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerStarted","Data":"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.284083 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerStarted","Data":"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.284161 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerStarted","Data":"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.284389 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-log" containerID="cri-o://8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" gracePeriod=30 Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.284559 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-metadata" containerID="cri-o://9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" gracePeriod=30 Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.295388 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0fca1e2c-1ea1-4239-aae1-6da063df9c3a","Type":"ContainerStarted","Data":"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47"} Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.311153 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" podStartSLOduration=6.311136943 podStartE2EDuration="6.311136943s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:20.280892533 +0000 UTC m=+1270.454205915" watchObservedRunningTime="2025-09-29 10:43:20.311136943 +0000 UTC m=+1270.484450305" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.319094 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.720448047 podStartE2EDuration="6.319071728s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="2025-09-29 10:43:15.521896195 +0000 UTC m=+1265.695209557" lastFinishedPulling="2025-09-29 10:43:19.120519876 +0000 UTC m=+1269.293833238" observedRunningTime="2025-09-29 10:43:20.306108194 +0000 UTC m=+1270.479421576" watchObservedRunningTime="2025-09-29 10:43:20.319071728 +0000 UTC m=+1270.492385090" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.335283 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.085824206 podStartE2EDuration="6.335261135s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="2025-09-29 10:43:15.856286786 +0000 UTC m=+1266.029600148" lastFinishedPulling="2025-09-29 10:43:19.105723725 +0000 UTC m=+1269.279037077" observedRunningTime="2025-09-29 10:43:20.329315572 +0000 UTC m=+1270.502628924" watchObservedRunningTime="2025-09-29 10:43:20.335261135 +0000 UTC m=+1270.508574497" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.367149 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.747568665 podStartE2EDuration="6.367126787s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="2025-09-29 10:43:15.488323679 +0000 UTC m=+1265.661637051" lastFinishedPulling="2025-09-29 10:43:19.107881811 +0000 UTC m=+1269.281195173" observedRunningTime="2025-09-29 10:43:20.348722032 +0000 UTC m=+1270.522035394" watchObservedRunningTime="2025-09-29 10:43:20.367126787 +0000 UTC m=+1270.540440159" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.375031 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.70662323 podStartE2EDuration="6.37501015s" podCreationTimestamp="2025-09-29 10:43:14 +0000 UTC" firstStartedPulling="2025-09-29 10:43:15.4390856 +0000 UTC m=+1265.612398952" lastFinishedPulling="2025-09-29 10:43:19.10747251 +0000 UTC m=+1269.280785872" observedRunningTime="2025-09-29 10:43:20.364199371 +0000 UTC m=+1270.537512753" watchObservedRunningTime="2025-09-29 10:43:20.37501015 +0000 UTC m=+1270.548323512" Sep 29 10:43:20 crc kubenswrapper[4727]: I0929 10:43:20.961814 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.076686 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data\") pod \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.076828 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle\") pod \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.076908 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs\") pod \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.076976 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvs6c\" (UniqueName: \"kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c\") pod \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\" (UID: \"9f0ed704-6a45-4eee-bb02-7228deb76eb0\") " Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.079766 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs" (OuterVolumeSpecName: "logs") pod "9f0ed704-6a45-4eee-bb02-7228deb76eb0" (UID: "9f0ed704-6a45-4eee-bb02-7228deb76eb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.109006 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c" (OuterVolumeSpecName: "kube-api-access-qvs6c") pod "9f0ed704-6a45-4eee-bb02-7228deb76eb0" (UID: "9f0ed704-6a45-4eee-bb02-7228deb76eb0"). InnerVolumeSpecName "kube-api-access-qvs6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.123449 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f0ed704-6a45-4eee-bb02-7228deb76eb0" (UID: "9f0ed704-6a45-4eee-bb02-7228deb76eb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.137173 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data" (OuterVolumeSpecName: "config-data") pod "9f0ed704-6a45-4eee-bb02-7228deb76eb0" (UID: "9f0ed704-6a45-4eee-bb02-7228deb76eb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.179717 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.179768 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0ed704-6a45-4eee-bb02-7228deb76eb0-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.179784 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvs6c\" (UniqueName: \"kubernetes.io/projected/9f0ed704-6a45-4eee-bb02-7228deb76eb0-kube-api-access-qvs6c\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.179798 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0ed704-6a45-4eee-bb02-7228deb76eb0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.264474 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.264783 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" containerName="kube-state-metrics" containerID="cri-o://b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34" gracePeriod=30 Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306413 4727 generic.go:334] "Generic (PLEG): container finished" podID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerID="9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" exitCode=0 Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306675 4727 generic.go:334] "Generic (PLEG): container finished" podID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerID="8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" exitCode=143 Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306485 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerDied","Data":"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3"} Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306727 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerDied","Data":"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7"} Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306749 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f0ed704-6a45-4eee-bb02-7228deb76eb0","Type":"ContainerDied","Data":"50ade81c394c316f0d0f5e6af22f0e540366547eac2033465a287746fa298648"} Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306769 4727 scope.go:117] "RemoveContainer" containerID="9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.306535 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.329812 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.338701 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.341115 4727 scope.go:117] "RemoveContainer" containerID="8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.367198 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:21 crc kubenswrapper[4727]: E0929 10:43:21.367591 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-metadata" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.367609 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-metadata" Sep 29 10:43:21 crc kubenswrapper[4727]: E0929 10:43:21.367623 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-log" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.367630 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-log" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.367848 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-metadata" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.367871 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" containerName="nova-metadata-log" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.368962 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.371548 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.371905 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.376769 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.444897 4727 scope.go:117] "RemoveContainer" containerID="9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" Sep 29 10:43:21 crc kubenswrapper[4727]: E0929 10:43:21.445378 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3\": container with ID starting with 9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3 not found: ID does not exist" containerID="9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.445409 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3"} err="failed to get container status \"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3\": rpc error: code = NotFound desc = could not find container \"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3\": container with ID starting with 9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3 not found: ID does not exist" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.445430 4727 scope.go:117] "RemoveContainer" containerID="8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" Sep 29 10:43:21 crc kubenswrapper[4727]: E0929 10:43:21.446214 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7\": container with ID starting with 8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7 not found: ID does not exist" containerID="8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.446329 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7"} err="failed to get container status \"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7\": rpc error: code = NotFound desc = could not find container \"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7\": container with ID starting with 8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7 not found: ID does not exist" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.446356 4727 scope.go:117] "RemoveContainer" containerID="9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.446707 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3"} err="failed to get container status \"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3\": rpc error: code = NotFound desc = could not find container \"9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3\": container with ID starting with 9ea7ad2760a0a5ccd0a410ac7c85dda21f6a681beecf5339e88c132666564fd3 not found: ID does not exist" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.446727 4727 scope.go:117] "RemoveContainer" containerID="8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.447179 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7"} err="failed to get container status \"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7\": rpc error: code = NotFound desc = could not find container \"8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7\": container with ID starting with 8da4c6860af71ab6caac770f22afd7cd0352044ead6d726fc01e68e2afcc97f7 not found: ID does not exist" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.502294 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.502464 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.502890 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5k2\" (UniqueName: \"kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.502937 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.502970 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.604662 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5k2\" (UniqueName: \"kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.604703 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.604740 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.604810 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.604876 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.605541 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.612085 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.612724 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.625218 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.632477 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5k2\" (UniqueName: \"kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2\") pod \"nova-metadata-0\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.767222 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.802805 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.910427 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zp8f\" (UniqueName: \"kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f\") pod \"03aebbc4-3f74-46f5-84b9-c0b7935569b7\" (UID: \"03aebbc4-3f74-46f5-84b9-c0b7935569b7\") " Sep 29 10:43:21 crc kubenswrapper[4727]: I0929 10:43:21.918620 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f" (OuterVolumeSpecName: "kube-api-access-4zp8f") pod "03aebbc4-3f74-46f5-84b9-c0b7935569b7" (UID: "03aebbc4-3f74-46f5-84b9-c0b7935569b7"). InnerVolumeSpecName "kube-api-access-4zp8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.013591 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zp8f\" (UniqueName: \"kubernetes.io/projected/03aebbc4-3f74-46f5-84b9-c0b7935569b7-kube-api-access-4zp8f\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.319602 4727 generic.go:334] "Generic (PLEG): container finished" podID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" containerID="b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34" exitCode=2 Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.319933 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03aebbc4-3f74-46f5-84b9-c0b7935569b7","Type":"ContainerDied","Data":"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34"} Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.319957 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03aebbc4-3f74-46f5-84b9-c0b7935569b7","Type":"ContainerDied","Data":"9684e84219ccd0abdc76e486a7541f648e284063a4970f063b2f930748080083"} Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.319972 4727 scope.go:117] "RemoveContainer" containerID="b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.320055 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.342704 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.355061 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.376740 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.377386 4727 scope.go:117] "RemoveContainer" containerID="b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34" Sep 29 10:43:22 crc kubenswrapper[4727]: E0929 10:43:22.377786 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34\": container with ID starting with b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34 not found: ID does not exist" containerID="b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.377824 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34"} err="failed to get container status \"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34\": rpc error: code = NotFound desc = could not find container \"b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34\": container with ID starting with b320e8ef569796112b20ae2180c82aaa5ede7b40a5772d247a742d200c00aa34 not found: ID does not exist" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.403284 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:22 crc kubenswrapper[4727]: E0929 10:43:22.403839 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" containerName="kube-state-metrics" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.403866 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" containerName="kube-state-metrics" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.404115 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" containerName="kube-state-metrics" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.405225 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.407698 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.410555 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.418979 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.521285 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.521407 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwvf\" (UniqueName: \"kubernetes.io/projected/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-api-access-dqwvf\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.521484 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.521651 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.623445 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.623513 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.623637 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.623720 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwvf\" (UniqueName: \"kubernetes.io/projected/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-api-access-dqwvf\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.629157 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.629325 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.629922 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.642593 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwvf\" (UniqueName: \"kubernetes.io/projected/bbc0072e-f90a-4518-8505-bee6f68abbaa-kube-api-access-dqwvf\") pod \"kube-state-metrics-0\" (UID: \"bbc0072e-f90a-4518-8505-bee6f68abbaa\") " pod="openstack/kube-state-metrics-0" Sep 29 10:43:22 crc kubenswrapper[4727]: I0929 10:43:22.730849 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.122767 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03aebbc4-3f74-46f5-84b9-c0b7935569b7" path="/var/lib/kubelet/pods/03aebbc4-3f74-46f5-84b9-c0b7935569b7/volumes" Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.123824 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0ed704-6a45-4eee-bb02-7228deb76eb0" path="/var/lib/kubelet/pods/9f0ed704-6a45-4eee-bb02-7228deb76eb0/volumes" Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.215276 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 29 10:43:23 crc kubenswrapper[4727]: W0929 10:43:23.226370 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbc0072e_f90a_4518_8505_bee6f68abbaa.slice/crio-f375c2afc4a9177dfa05952e19cd6509bec39212aa554e84125688dd97d97c77 WatchSource:0}: Error finding container f375c2afc4a9177dfa05952e19cd6509bec39212aa554e84125688dd97d97c77: Status 404 returned error can't find the container with id f375c2afc4a9177dfa05952e19cd6509bec39212aa554e84125688dd97d97c77 Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.286768 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.287284 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-central-agent" containerID="cri-o://48df589cd381bb4c8bb998173003118441f21802c8afc32012966fbdc41baa78" gracePeriod=30 Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.287371 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="sg-core" containerID="cri-o://d16983737a991b453b1cf8b8b0eaa67b15c90872c4017ec0f45b01377d0e9c53" gracePeriod=30 Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.287396 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="proxy-httpd" containerID="cri-o://3891805d5fdd737649f8ea893ba2b72448437cc39f2b0ae65726c379d8988405" gracePeriod=30 Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.287395 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-notification-agent" containerID="cri-o://71f325f1585194e04b800a7cc6972876d6bbea088a890589cc7ea0d55ef24492" gracePeriod=30 Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.381311 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbc0072e-f90a-4518-8505-bee6f68abbaa","Type":"ContainerStarted","Data":"f375c2afc4a9177dfa05952e19cd6509bec39212aa554e84125688dd97d97c77"} Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.383555 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerStarted","Data":"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc"} Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.383608 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerStarted","Data":"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b"} Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.383619 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerStarted","Data":"1e97b2a0a1d6c7deaa433c966a1fd67dbb6636af9a53eecf9961ffe47e66fdf4"} Sep 29 10:43:23 crc kubenswrapper[4727]: I0929 10:43:23.419704 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.41232179 podStartE2EDuration="2.41232179s" podCreationTimestamp="2025-09-29 10:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:23.406250214 +0000 UTC m=+1273.579563576" watchObservedRunningTime="2025-09-29 10:43:23.41232179 +0000 UTC m=+1273.585635152" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.394051 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbc0072e-f90a-4518-8505-bee6f68abbaa","Type":"ContainerStarted","Data":"0eeb13cfc07bf85b0a16471ba11f7cf57a8c09389ba7cee552e58a9e25f125de"} Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.394392 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396537 4727 generic.go:334] "Generic (PLEG): container finished" podID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerID="3891805d5fdd737649f8ea893ba2b72448437cc39f2b0ae65726c379d8988405" exitCode=0 Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396564 4727 generic.go:334] "Generic (PLEG): container finished" podID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerID="d16983737a991b453b1cf8b8b0eaa67b15c90872c4017ec0f45b01377d0e9c53" exitCode=2 Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396573 4727 generic.go:334] "Generic (PLEG): container finished" podID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerID="48df589cd381bb4c8bb998173003118441f21802c8afc32012966fbdc41baa78" exitCode=0 Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396600 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerDied","Data":"3891805d5fdd737649f8ea893ba2b72448437cc39f2b0ae65726c379d8988405"} Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396623 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerDied","Data":"d16983737a991b453b1cf8b8b0eaa67b15c90872c4017ec0f45b01377d0e9c53"} Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.396634 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerDied","Data":"48df589cd381bb4c8bb998173003118441f21802c8afc32012966fbdc41baa78"} Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.423261 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.879668411 podStartE2EDuration="2.423241605s" podCreationTimestamp="2025-09-29 10:43:22 +0000 UTC" firstStartedPulling="2025-09-29 10:43:23.230276878 +0000 UTC m=+1273.403590230" lastFinishedPulling="2025-09-29 10:43:23.773850062 +0000 UTC m=+1273.947163424" observedRunningTime="2025-09-29 10:43:24.411710557 +0000 UTC m=+1274.585023919" watchObservedRunningTime="2025-09-29 10:43:24.423241605 +0000 UTC m=+1274.596554967" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.834377 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.834716 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.865430 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.979621 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:43:24 crc kubenswrapper[4727]: I0929 10:43:24.979691 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.072497 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.085520 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.139905 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.140127 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="dnsmasq-dns" containerID="cri-o://c815a6292fd973c9eb2dc88428dbf3fcf81f7eaa88dc5f9db402c1ae9b7f899b" gracePeriod=10 Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.414623 4727 generic.go:334] "Generic (PLEG): container finished" podID="da99e317-fa23-4a56-a23d-1447b1ad3cb0" containerID="a34b5738bfed82f755ab116015123779cad6e88fe1eb7e97a94f7e491172bf58" exitCode=0 Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.414706 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dl7pr" event={"ID":"da99e317-fa23-4a56-a23d-1447b1ad3cb0","Type":"ContainerDied","Data":"a34b5738bfed82f755ab116015123779cad6e88fe1eb7e97a94f7e491172bf58"} Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.417395 4727 generic.go:334] "Generic (PLEG): container finished" podID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerID="c815a6292fd973c9eb2dc88428dbf3fcf81f7eaa88dc5f9db402c1ae9b7f899b" exitCode=0 Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.417476 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" event={"ID":"8150baa4-c7ed-4f99-9c0b-988c41630c97","Type":"ContainerDied","Data":"c815a6292fd973c9eb2dc88428dbf3fcf81f7eaa88dc5f9db402c1ae9b7f899b"} Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.466595 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.796314 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.923751 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.924620 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.924659 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.924715 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.924819 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhm5m\" (UniqueName: \"kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.924863 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config\") pod \"8150baa4-c7ed-4f99-9c0b-988c41630c97\" (UID: \"8150baa4-c7ed-4f99-9c0b-988c41630c97\") " Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.947583 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m" (OuterVolumeSpecName: "kube-api-access-nhm5m") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "kube-api-access-nhm5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.986175 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.992199 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.992370 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.993542 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config" (OuterVolumeSpecName: "config") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:43:25 crc kubenswrapper[4727]: I0929 10:43:25.998205 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8150baa4-c7ed-4f99-9c0b-988c41630c97" (UID: "8150baa4-c7ed-4f99-9c0b-988c41630c97"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026744 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026793 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhm5m\" (UniqueName: \"kubernetes.io/projected/8150baa4-c7ed-4f99-9c0b-988c41630c97-kube-api-access-nhm5m\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026813 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026825 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026836 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.026847 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8150baa4-c7ed-4f99-9c0b-988c41630c97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.061643 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.061865 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.429172 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" event={"ID":"8150baa4-c7ed-4f99-9c0b-988c41630c97","Type":"ContainerDied","Data":"14c29fd90f304ca9f3ba72ffd53e4f7a02107b1bfff7aefb5b5f5338690bac8c"} Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.429407 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bc8f75b9-xf5sx" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.429550 4727 scope.go:117] "RemoveContainer" containerID="c815a6292fd973c9eb2dc88428dbf3fcf81f7eaa88dc5f9db402c1ae9b7f899b" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.486499 4727 scope.go:117] "RemoveContainer" containerID="6f81c50977d0a38667578e1d0d80b7454bf7cad0b4ee7c87621786c0ab0f6680" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.511677 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.539365 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65bc8f75b9-xf5sx"] Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.769266 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.769319 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:43:26 crc kubenswrapper[4727]: I0929 10:43:26.913730 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.055572 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data\") pod \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.055789 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle\") pod \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.055869 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hdmr\" (UniqueName: \"kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr\") pod \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.055931 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts\") pod \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\" (UID: \"da99e317-fa23-4a56-a23d-1447b1ad3cb0\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.061585 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts" (OuterVolumeSpecName: "scripts") pod "da99e317-fa23-4a56-a23d-1447b1ad3cb0" (UID: "da99e317-fa23-4a56-a23d-1447b1ad3cb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.064508 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr" (OuterVolumeSpecName: "kube-api-access-6hdmr") pod "da99e317-fa23-4a56-a23d-1447b1ad3cb0" (UID: "da99e317-fa23-4a56-a23d-1447b1ad3cb0"). InnerVolumeSpecName "kube-api-access-6hdmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.090787 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da99e317-fa23-4a56-a23d-1447b1ad3cb0" (UID: "da99e317-fa23-4a56-a23d-1447b1ad3cb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.096639 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data" (OuterVolumeSpecName: "config-data") pod "da99e317-fa23-4a56-a23d-1447b1ad3cb0" (UID: "da99e317-fa23-4a56-a23d-1447b1ad3cb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.140828 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" path="/var/lib/kubelet/pods/8150baa4-c7ed-4f99-9c0b-988c41630c97/volumes" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.158230 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.158263 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.158278 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hdmr\" (UniqueName: \"kubernetes.io/projected/da99e317-fa23-4a56-a23d-1447b1ad3cb0-kube-api-access-6hdmr\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.158289 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da99e317-fa23-4a56-a23d-1447b1ad3cb0-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.442248 4727 generic.go:334] "Generic (PLEG): container finished" podID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerID="71f325f1585194e04b800a7cc6972876d6bbea088a890589cc7ea0d55ef24492" exitCode=0 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.444233 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerDied","Data":"71f325f1585194e04b800a7cc6972876d6bbea088a890589cc7ea0d55ef24492"} Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.444460 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.444466 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dl7pr" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.444471 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac6ad249-abc1-4de5-8210-ba66157bc5f6","Type":"ContainerDied","Data":"79d9beb297e188f258179b4dfd2d1af80fd0b11c4de216efd4aa1baccdf0f025"} Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.445025 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dl7pr" event={"ID":"da99e317-fa23-4a56-a23d-1447b1ad3cb0","Type":"ContainerDied","Data":"bf64ed6def6b24de87cb6b79886b86ea4be49e29c065c0e5ba7497b59f902ec1"} Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.445052 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf64ed6def6b24de87cb6b79886b86ea4be49e29c065c0e5ba7497b59f902ec1" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.445072 4727 scope.go:117] "RemoveContainer" containerID="3891805d5fdd737649f8ea893ba2b72448437cc39f2b0ae65726c379d8988405" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.477170 4727 scope.go:117] "RemoveContainer" containerID="d16983737a991b453b1cf8b8b0eaa67b15c90872c4017ec0f45b01377d0e9c53" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.525653 4727 scope.go:117] "RemoveContainer" containerID="71f325f1585194e04b800a7cc6972876d6bbea088a890589cc7ea0d55ef24492" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.566690 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.566807 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.566881 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.567005 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.567034 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.567157 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88bdz\" (UniqueName: \"kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.567189 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle\") pod \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\" (UID: \"ac6ad249-abc1-4de5-8210-ba66157bc5f6\") " Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.568470 4727 scope.go:117] "RemoveContainer" containerID="48df589cd381bb4c8bb998173003118441f21802c8afc32012966fbdc41baa78" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.570528 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.570883 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.574600 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz" (OuterVolumeSpecName: "kube-api-access-88bdz") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "kube-api-access-88bdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.599071 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts" (OuterVolumeSpecName: "scripts") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.611479 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.653134 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.653423 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-log" containerID="cri-o://70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c" gracePeriod=30 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.653838 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-api" containerID="cri-o://813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe" gracePeriod=30 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.677272 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.677461 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.677499 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerName="nova-scheduler-scheduler" containerID="cri-o://22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" gracePeriod=30 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.678016 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.678033 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.678044 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac6ad249-abc1-4de5-8210-ba66157bc5f6-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.678056 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88bdz\" (UniqueName: \"kubernetes.io/projected/ac6ad249-abc1-4de5-8210-ba66157bc5f6-kube-api-access-88bdz\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.711489 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.711721 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-log" containerID="cri-o://a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" gracePeriod=30 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.712109 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-metadata" containerID="cri-o://60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" gracePeriod=30 Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.753116 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.781143 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.834194 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data" (OuterVolumeSpecName: "config-data") pod "ac6ad249-abc1-4de5-8210-ba66157bc5f6" (UID: "ac6ad249-abc1-4de5-8210-ba66157bc5f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:27 crc kubenswrapper[4727]: I0929 10:43:27.885067 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac6ad249-abc1-4de5-8210-ba66157bc5f6-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.394851 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.457662 4727 generic.go:334] "Generic (PLEG): container finished" podID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerID="60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" exitCode=0 Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.458069 4727 generic.go:334] "Generic (PLEG): container finished" podID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerID="a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" exitCode=143 Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.457716 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.457733 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerDied","Data":"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc"} Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.458168 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerDied","Data":"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b"} Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.458184 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44cc42f1-ba27-40e4-a54e-609036f92a0b","Type":"ContainerDied","Data":"1e97b2a0a1d6c7deaa433c966a1fd67dbb6636af9a53eecf9961ffe47e66fdf4"} Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.458216 4727 scope.go:117] "RemoveContainer" containerID="60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.461784 4727 generic.go:334] "Generic (PLEG): container finished" podID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerID="70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c" exitCode=143 Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.461897 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerDied","Data":"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c"} Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.463245 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.495122 4727 scope.go:117] "RemoveContainer" containerID="a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.496548 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle\") pod \"44cc42f1-ba27-40e4-a54e-609036f92a0b\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.496627 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs\") pod \"44cc42f1-ba27-40e4-a54e-609036f92a0b\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.496817 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs\") pod \"44cc42f1-ba27-40e4-a54e-609036f92a0b\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.496850 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data\") pod \"44cc42f1-ba27-40e4-a54e-609036f92a0b\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.496950 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph5k2\" (UniqueName: \"kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2\") pod \"44cc42f1-ba27-40e4-a54e-609036f92a0b\" (UID: \"44cc42f1-ba27-40e4-a54e-609036f92a0b\") " Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.497318 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs" (OuterVolumeSpecName: "logs") pod "44cc42f1-ba27-40e4-a54e-609036f92a0b" (UID: "44cc42f1-ba27-40e4-a54e-609036f92a0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.498163 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44cc42f1-ba27-40e4-a54e-609036f92a0b-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.502945 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2" (OuterVolumeSpecName: "kube-api-access-ph5k2") pod "44cc42f1-ba27-40e4-a54e-609036f92a0b" (UID: "44cc42f1-ba27-40e4-a54e-609036f92a0b"). InnerVolumeSpecName "kube-api-access-ph5k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.508414 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.519646 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.534694 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535150 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="init" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535167 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="init" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535181 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-central-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535188 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-central-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535199 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-notification-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535206 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-notification-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535220 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="proxy-httpd" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535226 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="proxy-httpd" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535237 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="sg-core" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535243 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="sg-core" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535254 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="dnsmasq-dns" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535259 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="dnsmasq-dns" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535278 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-metadata" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535285 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-metadata" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535297 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-log" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535303 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-log" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.535315 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da99e317-fa23-4a56-a23d-1447b1ad3cb0" containerName="nova-manage" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535320 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="da99e317-fa23-4a56-a23d-1447b1ad3cb0" containerName="nova-manage" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535530 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="proxy-httpd" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535547 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="da99e317-fa23-4a56-a23d-1447b1ad3cb0" containerName="nova-manage" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535561 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-central-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535577 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="ceilometer-notification-agent" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535589 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="8150baa4-c7ed-4f99-9c0b-988c41630c97" containerName="dnsmasq-dns" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535598 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" containerName="sg-core" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535613 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-log" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.535624 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" containerName="nova-metadata-metadata" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.537877 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.541422 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44cc42f1-ba27-40e4-a54e-609036f92a0b" (UID: "44cc42f1-ba27-40e4-a54e-609036f92a0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.542055 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.543409 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.544123 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.552584 4727 scope.go:117] "RemoveContainer" containerID="60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.563469 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc\": container with ID starting with 60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc not found: ID does not exist" containerID="60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.563518 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc"} err="failed to get container status \"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc\": rpc error: code = NotFound desc = could not find container \"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc\": container with ID starting with 60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc not found: ID does not exist" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.563545 4727 scope.go:117] "RemoveContainer" containerID="a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" Sep 29 10:43:28 crc kubenswrapper[4727]: E0929 10:43:28.564475 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b\": container with ID starting with a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b not found: ID does not exist" containerID="a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.564520 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b"} err="failed to get container status \"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b\": rpc error: code = NotFound desc = could not find container \"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b\": container with ID starting with a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b not found: ID does not exist" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.564552 4727 scope.go:117] "RemoveContainer" containerID="60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.564892 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc"} err="failed to get container status \"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc\": rpc error: code = NotFound desc = could not find container \"60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc\": container with ID starting with 60b90a931561ef28e07980f55f976abd40157bafeea93a23326b3673286c25fc not found: ID does not exist" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.564934 4727 scope.go:117] "RemoveContainer" containerID="a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.565219 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b"} err="failed to get container status \"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b\": rpc error: code = NotFound desc = could not find container \"a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b\": container with ID starting with a82132a397f1a88ff68a0f7f6c51e889664d927f5299cbb5a949b1ded3af7f4b not found: ID does not exist" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.567836 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.569415 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data" (OuterVolumeSpecName: "config-data") pod "44cc42f1-ba27-40e4-a54e-609036f92a0b" (UID: "44cc42f1-ba27-40e4-a54e-609036f92a0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.582742 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "44cc42f1-ba27-40e4-a54e-609036f92a0b" (UID: "44cc42f1-ba27-40e4-a54e-609036f92a0b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.600247 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.600277 4727 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.600287 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44cc42f1-ba27-40e4-a54e-609036f92a0b-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.600297 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph5k2\" (UniqueName: \"kubernetes.io/projected/44cc42f1-ba27-40e4-a54e-609036f92a0b-kube-api-access-ph5k2\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701752 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701842 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701869 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701904 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701923 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701942 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.701975 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7mj8\" (UniqueName: \"kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.702002 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.800042 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803449 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803493 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803515 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803551 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7mj8\" (UniqueName: \"kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803581 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803635 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.803900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.804158 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.804179 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.804325 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.807072 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.808913 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.808922 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.809165 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.812883 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.815534 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.827796 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.829734 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.830111 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7mj8\" (UniqueName: \"kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8\") pod \"ceilometer-0\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " pod="openstack/ceilometer-0" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.832599 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.832782 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.861713 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:28 crc kubenswrapper[4727]: I0929 10:43:28.969579 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.009815 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98mls\" (UniqueName: \"kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.009882 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.010136 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.010399 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.010453 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.137053 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98mls\" (UniqueName: \"kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.137447 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.137507 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.137559 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.137576 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.139124 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.142677 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.142900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.143818 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.154762 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44cc42f1-ba27-40e4-a54e-609036f92a0b" path="/var/lib/kubelet/pods/44cc42f1-ba27-40e4-a54e-609036f92a0b/volumes" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.155800 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac6ad249-abc1-4de5-8210-ba66157bc5f6" path="/var/lib/kubelet/pods/ac6ad249-abc1-4de5-8210-ba66157bc5f6/volumes" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.157493 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98mls\" (UniqueName: \"kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls\") pod \"nova-metadata-0\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.446185 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.500032 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:29 crc kubenswrapper[4727]: W0929 10:43:29.505041 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod613114a6_4186_4c9d_ba7c_f5d7029006ab.slice/crio-335a735f66295d2cf72cf0aae4b3f01a2c94d13a0ae09713263594a54e08c1c2 WatchSource:0}: Error finding container 335a735f66295d2cf72cf0aae4b3f01a2c94d13a0ae09713263594a54e08c1c2: Status 404 returned error can't find the container with id 335a735f66295d2cf72cf0aae4b3f01a2c94d13a0ae09713263594a54e08c1c2 Sep 29 10:43:29 crc kubenswrapper[4727]: E0929 10:43:29.835408 4727 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 10:43:29 crc kubenswrapper[4727]: E0929 10:43:29.836959 4727 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 10:43:29 crc kubenswrapper[4727]: E0929 10:43:29.838055 4727 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 29 10:43:29 crc kubenswrapper[4727]: E0929 10:43:29.838108 4727 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerName="nova-scheduler-scheduler" Sep 29 10:43:29 crc kubenswrapper[4727]: I0929 10:43:29.892442 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:43:29 crc kubenswrapper[4727]: W0929 10:43:29.901893 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c2a19c0_67ab_4406_8a75_e18012681d9d.slice/crio-d1835b28eb65e2df607f8fa2e3ef1453cc0c9640f4e4104f3c685d9e82f9b0ce WatchSource:0}: Error finding container d1835b28eb65e2df607f8fa2e3ef1453cc0c9640f4e4104f3c685d9e82f9b0ce: Status 404 returned error can't find the container with id d1835b28eb65e2df607f8fa2e3ef1453cc0c9640f4e4104f3c685d9e82f9b0ce Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.486981 4727 generic.go:334] "Generic (PLEG): container finished" podID="6027aeb9-0794-453e-85e3-9bf0e0c35d61" containerID="4cd2c1e0dcfb72c0d3f22f6f8d299f86231c1885b308943dd7207aea3df69e55" exitCode=0 Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.487046 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" event={"ID":"6027aeb9-0794-453e-85e3-9bf0e0c35d61","Type":"ContainerDied","Data":"4cd2c1e0dcfb72c0d3f22f6f8d299f86231c1885b308943dd7207aea3df69e55"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.489811 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerStarted","Data":"78bb6ff6d9c67306b0f27f5f7eacc5652e8ed2b48a081d56a949798e4a3112df"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.489839 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerStarted","Data":"335a735f66295d2cf72cf0aae4b3f01a2c94d13a0ae09713263594a54e08c1c2"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.492396 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerStarted","Data":"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.492422 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerStarted","Data":"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.492432 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerStarted","Data":"d1835b28eb65e2df607f8fa2e3ef1453cc0c9640f4e4104f3c685d9e82f9b0ce"} Sep 29 10:43:30 crc kubenswrapper[4727]: I0929 10:43:30.521816 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.52180053 podStartE2EDuration="2.52180053s" podCreationTimestamp="2025-09-29 10:43:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:30.519280985 +0000 UTC m=+1280.692594347" watchObservedRunningTime="2025-09-29 10:43:30.52180053 +0000 UTC m=+1280.695113892" Sep 29 10:43:31 crc kubenswrapper[4727]: I0929 10:43:31.504128 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerStarted","Data":"f15f2a5e4ae8fca3078ed7192266eefe47a5bf68649866573a3761ddb8b558ae"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.021452 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.152081 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.204203 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts\") pod \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.204427 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle\") pod \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.204504 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlt9c\" (UniqueName: \"kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c\") pod \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.204544 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data\") pod \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\" (UID: \"6027aeb9-0794-453e-85e3-9bf0e0c35d61\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.219846 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c" (OuterVolumeSpecName: "kube-api-access-rlt9c") pod "6027aeb9-0794-453e-85e3-9bf0e0c35d61" (UID: "6027aeb9-0794-453e-85e3-9bf0e0c35d61"). InnerVolumeSpecName "kube-api-access-rlt9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.225132 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts" (OuterVolumeSpecName: "scripts") pod "6027aeb9-0794-453e-85e3-9bf0e0c35d61" (UID: "6027aeb9-0794-453e-85e3-9bf0e0c35d61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.240837 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6027aeb9-0794-453e-85e3-9bf0e0c35d61" (UID: "6027aeb9-0794-453e-85e3-9bf0e0c35d61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.247282 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data" (OuterVolumeSpecName: "config-data") pod "6027aeb9-0794-453e-85e3-9bf0e0c35d61" (UID: "6027aeb9-0794-453e-85e3-9bf0e0c35d61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.306466 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle\") pod \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.306632 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hvtp\" (UniqueName: \"kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp\") pod \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.306809 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data\") pod \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\" (UID: \"0fca1e2c-1ea1-4239-aae1-6da063df9c3a\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.307463 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.307489 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.307504 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlt9c\" (UniqueName: \"kubernetes.io/projected/6027aeb9-0794-453e-85e3-9bf0e0c35d61-kube-api-access-rlt9c\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.307518 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6027aeb9-0794-453e-85e3-9bf0e0c35d61-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.313489 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp" (OuterVolumeSpecName: "kube-api-access-4hvtp") pod "0fca1e2c-1ea1-4239-aae1-6da063df9c3a" (UID: "0fca1e2c-1ea1-4239-aae1-6da063df9c3a"). InnerVolumeSpecName "kube-api-access-4hvtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.350630 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data" (OuterVolumeSpecName: "config-data") pod "0fca1e2c-1ea1-4239-aae1-6da063df9c3a" (UID: "0fca1e2c-1ea1-4239-aae1-6da063df9c3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.354588 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fca1e2c-1ea1-4239-aae1-6da063df9c3a" (UID: "0fca1e2c-1ea1-4239-aae1-6da063df9c3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.410394 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.410445 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hvtp\" (UniqueName: \"kubernetes.io/projected/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-kube-api-access-4hvtp\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.410462 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca1e2c-1ea1-4239-aae1-6da063df9c3a-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.506266 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.521849 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" event={"ID":"6027aeb9-0794-453e-85e3-9bf0e0c35d61","Type":"ContainerDied","Data":"cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.521919 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cadbf99f5a8b965c6db7f7f25da49aca9748c5149b7f19e029b0c9089ec7693c" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.522018 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8n5tt" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.536403 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerStarted","Data":"bc918bdbcb9b4f0ec39620940254dc61aed8d8a0c07d680c4b30c0bf0777bf3c"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.546679 4727 generic.go:334] "Generic (PLEG): container finished" podID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerID="813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe" exitCode=0 Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.546784 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerDied","Data":"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.546799 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.546823 4727 scope.go:117] "RemoveContainer" containerID="813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.546811 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13e791f8-a90a-4d87-934a-9bb569c9d6a7","Type":"ContainerDied","Data":"edc4da5a6b93578415f2eacc2904a322e7e4ec3246d572a05764e88a6f358092"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.552540 4727 generic.go:334] "Generic (PLEG): container finished" podID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" exitCode=0 Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.552584 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0fca1e2c-1ea1-4239-aae1-6da063df9c3a","Type":"ContainerDied","Data":"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.552607 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0fca1e2c-1ea1-4239-aae1-6da063df9c3a","Type":"ContainerDied","Data":"3f7e9683ee4f1afde98bbc60f4121b6a6e16fc33d69c0b340d881c8c50df69db"} Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.552589 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.585458 4727 scope.go:117] "RemoveContainer" containerID="70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606226 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.606664 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-api" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606685 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-api" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.606699 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-log" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606707 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-log" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.606731 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerName="nova-scheduler-scheduler" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606738 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerName="nova-scheduler-scheduler" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.606753 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6027aeb9-0794-453e-85e3-9bf0e0c35d61" containerName="nova-cell1-conductor-db-sync" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606759 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="6027aeb9-0794-453e-85e3-9bf0e0c35d61" containerName="nova-cell1-conductor-db-sync" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606928 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-api" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606949 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" containerName="nova-api-log" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606960 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" containerName="nova-scheduler-scheduler" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.606980 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="6027aeb9-0794-453e-85e3-9bf0e0c35d61" containerName="nova-cell1-conductor-db-sync" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.607594 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.609229 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.612377 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrhf9\" (UniqueName: \"kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9\") pod \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.612506 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data\") pod \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.612625 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs\") pod \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.612719 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle\") pod \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\" (UID: \"13e791f8-a90a-4d87-934a-9bb569c9d6a7\") " Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.613574 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs" (OuterVolumeSpecName: "logs") pod "13e791f8-a90a-4d87-934a-9bb569c9d6a7" (UID: "13e791f8-a90a-4d87-934a-9bb569c9d6a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.617102 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9" (OuterVolumeSpecName: "kube-api-access-xrhf9") pod "13e791f8-a90a-4d87-934a-9bb569c9d6a7" (UID: "13e791f8-a90a-4d87-934a-9bb569c9d6a7"). InnerVolumeSpecName "kube-api-access-xrhf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.626083 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.654901 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data" (OuterVolumeSpecName: "config-data") pod "13e791f8-a90a-4d87-934a-9bb569c9d6a7" (UID: "13e791f8-a90a-4d87-934a-9bb569c9d6a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.662408 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13e791f8-a90a-4d87-934a-9bb569c9d6a7" (UID: "13e791f8-a90a-4d87-934a-9bb569c9d6a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.714960 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715061 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmhn2\" (UniqueName: \"kubernetes.io/projected/2522d266-4b11-4a05-96fc-69d5847a63c0-kube-api-access-fmhn2\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715082 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715220 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrhf9\" (UniqueName: \"kubernetes.io/projected/13e791f8-a90a-4d87-934a-9bb569c9d6a7-kube-api-access-xrhf9\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715231 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715240 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13e791f8-a90a-4d87-934a-9bb569c9d6a7-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.715250 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e791f8-a90a-4d87-934a-9bb569c9d6a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.739606 4727 scope.go:117] "RemoveContainer" containerID="813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.740032 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe\": container with ID starting with 813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe not found: ID does not exist" containerID="813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.740082 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe"} err="failed to get container status \"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe\": rpc error: code = NotFound desc = could not find container \"813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe\": container with ID starting with 813504faa1b2a6f73fb6199f842c17ceef06510b1813c91af23bf66204ff35fe not found: ID does not exist" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.740110 4727 scope.go:117] "RemoveContainer" containerID="70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.740434 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c\": container with ID starting with 70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c not found: ID does not exist" containerID="70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.740457 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c"} err="failed to get container status \"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c\": rpc error: code = NotFound desc = could not find container \"70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c\": container with ID starting with 70dab9ed46fbb2e05762d8060763412d61b413e4dd9c7a2389c9cb7fd11e489c not found: ID does not exist" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.740473 4727 scope.go:117] "RemoveContainer" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.753786 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.762887 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.773669 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.777163 4727 scope.go:117] "RemoveContainer" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" Sep 29 10:43:32 crc kubenswrapper[4727]: E0929 10:43:32.780202 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47\": container with ID starting with 22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47 not found: ID does not exist" containerID="22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.780243 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47"} err="failed to get container status \"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47\": rpc error: code = NotFound desc = could not find container \"22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47\": container with ID starting with 22c652f1ab4cc3cf7ca11df100641e6f4c6bccec2a61109bef7da63593677a47 not found: ID does not exist" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.802506 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.803790 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.806638 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.819383 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmhn2\" (UniqueName: \"kubernetes.io/projected/2522d266-4b11-4a05-96fc-69d5847a63c0-kube-api-access-fmhn2\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.819425 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.819556 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.825282 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.826110 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522d266-4b11-4a05-96fc-69d5847a63c0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.826173 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.834911 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmhn2\" (UniqueName: \"kubernetes.io/projected/2522d266-4b11-4a05-96fc-69d5847a63c0-kube-api-access-fmhn2\") pod \"nova-cell1-conductor-0\" (UID: \"2522d266-4b11-4a05-96fc-69d5847a63c0\") " pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.917364 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.921161 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.921316 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn9sj\" (UniqueName: \"kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.921439 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.923533 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.937847 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.939660 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.943107 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 10:43:32 crc kubenswrapper[4727]: I0929 10:43:32.954146 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.023116 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.023221 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn9sj\" (UniqueName: \"kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.023575 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.027600 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.027851 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.039003 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn9sj\" (UniqueName: \"kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj\") pod \"nova-scheduler-0\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.057789 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.124583 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fca1e2c-1ea1-4239-aae1-6da063df9c3a" path="/var/lib/kubelet/pods/0fca1e2c-1ea1-4239-aae1-6da063df9c3a/volumes" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.125124 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e791f8-a90a-4d87-934a-9bb569c9d6a7" path="/var/lib/kubelet/pods/13e791f8-a90a-4d87-934a-9bb569c9d6a7/volumes" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.125996 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.126633 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.126823 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m8p6\" (UniqueName: \"kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.126873 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.127162 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.230829 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.231177 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m8p6\" (UniqueName: \"kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.231222 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.231325 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.232124 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.238330 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.256559 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m8p6\" (UniqueName: \"kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.258914 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data\") pod \"nova-api-0\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.265514 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:43:33 crc kubenswrapper[4727]: W0929 10:43:33.564162 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2522d266_4b11_4a05_96fc_69d5847a63c0.slice/crio-174e4ebba71911582484737584dc87dd74ee318343fca90b9f9c5cdc8f5726e0 WatchSource:0}: Error finding container 174e4ebba71911582484737584dc87dd74ee318343fca90b9f9c5cdc8f5726e0: Status 404 returned error can't find the container with id 174e4ebba71911582484737584dc87dd74ee318343fca90b9f9c5cdc8f5726e0 Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.566696 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.589694 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerStarted","Data":"a5f59d1c429ff590de47c6d320f27917f0645433ee57cad83086b9161566c13d"} Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.591807 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.618318 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.065946569 podStartE2EDuration="5.618301717s" podCreationTimestamp="2025-09-29 10:43:28 +0000 UTC" firstStartedPulling="2025-09-29 10:43:29.507998883 +0000 UTC m=+1279.681312255" lastFinishedPulling="2025-09-29 10:43:33.060354041 +0000 UTC m=+1283.233667403" observedRunningTime="2025-09-29 10:43:33.609965432 +0000 UTC m=+1283.783278794" watchObservedRunningTime="2025-09-29 10:43:33.618301717 +0000 UTC m=+1283.791615079" Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.709403 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:43:33 crc kubenswrapper[4727]: W0929 10:43:33.728444 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0db3ea4_126f_4c58_bbd4_859916d5347f.slice/crio-4985f1c4ce491b10e43282430247dda2259970f4bf65736732aa17bd44dfb264 WatchSource:0}: Error finding container 4985f1c4ce491b10e43282430247dda2259970f4bf65736732aa17bd44dfb264: Status 404 returned error can't find the container with id 4985f1c4ce491b10e43282430247dda2259970f4bf65736732aa17bd44dfb264 Sep 29 10:43:33 crc kubenswrapper[4727]: I0929 10:43:33.781817 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.447069 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.447477 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.600119 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerStarted","Data":"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.601384 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerStarted","Data":"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.601492 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerStarted","Data":"41071cbbad8659bc5e281e98c99303979f351a64b841029237b481d94ecca807"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.604430 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0db3ea4-126f-4c58-bbd4-859916d5347f","Type":"ContainerStarted","Data":"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.604584 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0db3ea4-126f-4c58-bbd4-859916d5347f","Type":"ContainerStarted","Data":"4985f1c4ce491b10e43282430247dda2259970f4bf65736732aa17bd44dfb264"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.607795 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2522d266-4b11-4a05-96fc-69d5847a63c0","Type":"ContainerStarted","Data":"28349c5d24ab787d54093de68f1bfe1d83713613097efd3d46fc3795637be62f"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.607930 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2522d266-4b11-4a05-96fc-69d5847a63c0","Type":"ContainerStarted","Data":"174e4ebba71911582484737584dc87dd74ee318343fca90b9f9c5cdc8f5726e0"} Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.607996 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.620972 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.620954027 podStartE2EDuration="2.620954027s" podCreationTimestamp="2025-09-29 10:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:34.61799215 +0000 UTC m=+1284.791305512" watchObservedRunningTime="2025-09-29 10:43:34.620954027 +0000 UTC m=+1284.794267389" Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.645744 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.645722215 podStartE2EDuration="2.645722215s" podCreationTimestamp="2025-09-29 10:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:34.636126488 +0000 UTC m=+1284.809439850" watchObservedRunningTime="2025-09-29 10:43:34.645722215 +0000 UTC m=+1284.819035597" Sep 29 10:43:34 crc kubenswrapper[4727]: I0929 10:43:34.662809 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.662787175 podStartE2EDuration="2.662787175s" podCreationTimestamp="2025-09-29 10:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:34.653326651 +0000 UTC m=+1284.826640013" watchObservedRunningTime="2025-09-29 10:43:34.662787175 +0000 UTC m=+1284.836100537" Sep 29 10:43:38 crc kubenswrapper[4727]: I0929 10:43:38.097223 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 29 10:43:38 crc kubenswrapper[4727]: I0929 10:43:38.126329 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 10:43:39 crc kubenswrapper[4727]: I0929 10:43:39.447077 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 10:43:39 crc kubenswrapper[4727]: I0929 10:43:39.447134 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 10:43:40 crc kubenswrapper[4727]: I0929 10:43:40.464538 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:40 crc kubenswrapper[4727]: I0929 10:43:40.464626 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:43 crc kubenswrapper[4727]: I0929 10:43:43.126343 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 10:43:43 crc kubenswrapper[4727]: I0929 10:43:43.158657 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 10:43:43 crc kubenswrapper[4727]: I0929 10:43:43.266495 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:43:43 crc kubenswrapper[4727]: I0929 10:43:43.266550 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:43:43 crc kubenswrapper[4727]: I0929 10:43:43.752412 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 10:43:44 crc kubenswrapper[4727]: I0929 10:43:44.348539 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:44 crc kubenswrapper[4727]: I0929 10:43:44.348539 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.247397 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.247828 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.451956 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.452636 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.459277 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 10:43:49 crc kubenswrapper[4727]: I0929 10:43:49.783823 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.632783 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.739386 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfbgb\" (UniqueName: \"kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb\") pod \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.739455 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle\") pod \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.739516 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data\") pod \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\" (UID: \"a1232ec7-a034-4ed1-845b-0ffc5bb900d7\") " Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.745249 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb" (OuterVolumeSpecName: "kube-api-access-zfbgb") pod "a1232ec7-a034-4ed1-845b-0ffc5bb900d7" (UID: "a1232ec7-a034-4ed1-845b-0ffc5bb900d7"). InnerVolumeSpecName "kube-api-access-zfbgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.770248 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data" (OuterVolumeSpecName: "config-data") pod "a1232ec7-a034-4ed1-845b-0ffc5bb900d7" (UID: "a1232ec7-a034-4ed1-845b-0ffc5bb900d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.790666 4727 generic.go:334] "Generic (PLEG): container finished" podID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" containerID="b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349" exitCode=137 Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.790989 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1232ec7-a034-4ed1-845b-0ffc5bb900d7" (UID: "a1232ec7-a034-4ed1-845b-0ffc5bb900d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.791545 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.791827 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1232ec7-a034-4ed1-845b-0ffc5bb900d7","Type":"ContainerDied","Data":"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349"} Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.791859 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1232ec7-a034-4ed1-845b-0ffc5bb900d7","Type":"ContainerDied","Data":"424d1fcd8323664b1ff76a9cf65a9ca920321c2b85245c80896d52cf3c9cb835"} Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.791877 4727 scope.go:117] "RemoveContainer" containerID="b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.841801 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfbgb\" (UniqueName: \"kubernetes.io/projected/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-kube-api-access-zfbgb\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.841834 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.841847 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1232ec7-a034-4ed1-845b-0ffc5bb900d7-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.865476 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.870445 4727 scope.go:117] "RemoveContainer" containerID="b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349" Sep 29 10:43:50 crc kubenswrapper[4727]: E0929 10:43:50.870940 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349\": container with ID starting with b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349 not found: ID does not exist" containerID="b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.870983 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349"} err="failed to get container status \"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349\": rpc error: code = NotFound desc = could not find container \"b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349\": container with ID starting with b371eb4e368987580825b8f8a197fd587d7d488672f080c25f63e953ca3d5349 not found: ID does not exist" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.876880 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.888555 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:50 crc kubenswrapper[4727]: E0929 10:43:50.889103 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.889127 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.889359 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" containerName="nova-cell1-novncproxy-novncproxy" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.890173 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.891863 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.892182 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.899120 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 29 10:43:50 crc kubenswrapper[4727]: I0929 10:43:50.904383 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.044603 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.045028 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p79sj\" (UniqueName: \"kubernetes.io/projected/6267f772-555e-4c6f-b937-6d8a4c61dab3-kube-api-access-p79sj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.045136 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.045287 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.045412 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.124731 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1232ec7-a034-4ed1-845b-0ffc5bb900d7" path="/var/lib/kubelet/pods/a1232ec7-a034-4ed1-845b-0ffc5bb900d7/volumes" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.147989 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p79sj\" (UniqueName: \"kubernetes.io/projected/6267f772-555e-4c6f-b937-6d8a4c61dab3-kube-api-access-p79sj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.148089 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.148195 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.148231 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.148261 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.154202 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.154525 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.154759 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.155035 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6267f772-555e-4c6f-b937-6d8a4c61dab3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.170260 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p79sj\" (UniqueName: \"kubernetes.io/projected/6267f772-555e-4c6f-b937-6d8a4c61dab3-kube-api-access-p79sj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6267f772-555e-4c6f-b937-6d8a4c61dab3\") " pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.246285 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:51 crc kubenswrapper[4727]: W0929 10:43:51.674206 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6267f772_555e_4c6f_b937_6d8a4c61dab3.slice/crio-778ec5a09b0457a7096aac8bae472562ce0630900ce2e63595845a2f28bcd5d7 WatchSource:0}: Error finding container 778ec5a09b0457a7096aac8bae472562ce0630900ce2e63595845a2f28bcd5d7: Status 404 returned error can't find the container with id 778ec5a09b0457a7096aac8bae472562ce0630900ce2e63595845a2f28bcd5d7 Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.682719 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 29 10:43:51 crc kubenswrapper[4727]: I0929 10:43:51.801709 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6267f772-555e-4c6f-b937-6d8a4c61dab3","Type":"ContainerStarted","Data":"778ec5a09b0457a7096aac8bae472562ce0630900ce2e63595845a2f28bcd5d7"} Sep 29 10:43:52 crc kubenswrapper[4727]: I0929 10:43:52.811764 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6267f772-555e-4c6f-b937-6d8a4c61dab3","Type":"ContainerStarted","Data":"b032ab6bcc2fb118dafb36dfbea462bf34427a155738a3b938e5694607cd3618"} Sep 29 10:43:52 crc kubenswrapper[4727]: I0929 10:43:52.833022 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.832989428 podStartE2EDuration="2.832989428s" podCreationTimestamp="2025-09-29 10:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:52.830018661 +0000 UTC m=+1303.003332033" watchObservedRunningTime="2025-09-29 10:43:52.832989428 +0000 UTC m=+1303.006302830" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.270627 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.271256 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.271535 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.275883 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.821689 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 10:43:53 crc kubenswrapper[4727]: I0929 10:43:53.920978 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.083176 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.094250 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.112851 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.220088 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.220168 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.220199 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.220287 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.221021 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.221073 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss6zd\" (UniqueName: \"kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.322314 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.322528 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.322696 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.322878 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.323015 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss6zd\" (UniqueName: \"kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.323150 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.323408 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.323715 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.323890 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.324310 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.324482 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.346015 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss6zd\" (UniqueName: \"kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd\") pod \"dnsmasq-dns-85649f948c-cbw9p\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.421032 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:54 crc kubenswrapper[4727]: I0929 10:43:54.913987 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:43:55 crc kubenswrapper[4727]: I0929 10:43:55.851461 4727 generic.go:334] "Generic (PLEG): container finished" podID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerID="2277155b0c7fb8ae2aa23f083f62ecae16d7e68f4feaaa5386b0bf4be2b7cb5a" exitCode=0 Sep 29 10:43:55 crc kubenswrapper[4727]: I0929 10:43:55.853173 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" event={"ID":"b0e563a1-166c-4155-a5d9-35c5272c089f","Type":"ContainerDied","Data":"2277155b0c7fb8ae2aa23f083f62ecae16d7e68f4feaaa5386b0bf4be2b7cb5a"} Sep 29 10:43:55 crc kubenswrapper[4727]: I0929 10:43:55.853213 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" event={"ID":"b0e563a1-166c-4155-a5d9-35c5272c089f","Type":"ContainerStarted","Data":"e7c080106de0c9dfb8a930e1e927bcf9d1f4b11a50e26b89be06497498612431"} Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.131100 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.131388 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-central-agent" containerID="cri-o://78bb6ff6d9c67306b0f27f5f7eacc5652e8ed2b48a081d56a949798e4a3112df" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.131424 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="proxy-httpd" containerID="cri-o://a5f59d1c429ff590de47c6d320f27917f0645433ee57cad83086b9161566c13d" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.131508 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="sg-core" containerID="cri-o://bc918bdbcb9b4f0ec39620940254dc61aed8d8a0c07d680c4b30c0bf0777bf3c" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.131554 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-notification-agent" containerID="cri-o://f15f2a5e4ae8fca3078ed7192266eefe47a5bf68649866573a3761ddb8b558ae" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.145282 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.197:3000/\": EOF" Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.247159 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.493600 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.864900 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" event={"ID":"b0e563a1-166c-4155-a5d9-35c5272c089f","Type":"ContainerStarted","Data":"d65da3163f083133f75b9a9838103e80ab44ecb5001e8c74a8e416d28cf569e9"} Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.865039 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867670 4727 generic.go:334] "Generic (PLEG): container finished" podID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerID="a5f59d1c429ff590de47c6d320f27917f0645433ee57cad83086b9161566c13d" exitCode=0 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867696 4727 generic.go:334] "Generic (PLEG): container finished" podID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerID="bc918bdbcb9b4f0ec39620940254dc61aed8d8a0c07d680c4b30c0bf0777bf3c" exitCode=2 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867717 4727 generic.go:334] "Generic (PLEG): container finished" podID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerID="78bb6ff6d9c67306b0f27f5f7eacc5652e8ed2b48a081d56a949798e4a3112df" exitCode=0 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867736 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerDied","Data":"a5f59d1c429ff590de47c6d320f27917f0645433ee57cad83086b9161566c13d"} Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867757 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerDied","Data":"bc918bdbcb9b4f0ec39620940254dc61aed8d8a0c07d680c4b30c0bf0777bf3c"} Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867769 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerDied","Data":"78bb6ff6d9c67306b0f27f5f7eacc5652e8ed2b48a081d56a949798e4a3112df"} Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867906 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-log" containerID="cri-o://970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.867945 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-api" containerID="cri-o://d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5" gracePeriod=30 Sep 29 10:43:56 crc kubenswrapper[4727]: I0929 10:43:56.898585 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" podStartSLOduration=2.898565289 podStartE2EDuration="2.898565289s" podCreationTimestamp="2025-09-29 10:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:43:56.892373419 +0000 UTC m=+1307.065686801" watchObservedRunningTime="2025-09-29 10:43:56.898565289 +0000 UTC m=+1307.071878651" Sep 29 10:43:57 crc kubenswrapper[4727]: I0929 10:43:57.884023 4727 generic.go:334] "Generic (PLEG): container finished" podID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerID="f15f2a5e4ae8fca3078ed7192266eefe47a5bf68649866573a3761ddb8b558ae" exitCode=0 Sep 29 10:43:57 crc kubenswrapper[4727]: I0929 10:43:57.884116 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerDied","Data":"f15f2a5e4ae8fca3078ed7192266eefe47a5bf68649866573a3761ddb8b558ae"} Sep 29 10:43:57 crc kubenswrapper[4727]: I0929 10:43:57.886827 4727 generic.go:334] "Generic (PLEG): container finished" podID="11290f51-f732-4756-8e1e-701a8699464f" containerID="970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4" exitCode=143 Sep 29 10:43:57 crc kubenswrapper[4727]: I0929 10:43:57.886902 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerDied","Data":"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4"} Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.182637 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244025 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244216 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244278 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244320 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244378 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244450 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7mj8\" (UniqueName: \"kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244481 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244519 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts\") pod \"613114a6-4186-4c9d-ba7c-f5d7029006ab\" (UID: \"613114a6-4186-4c9d-ba7c-f5d7029006ab\") " Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244840 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.245378 4727 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.244869 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.262026 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts" (OuterVolumeSpecName: "scripts") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.262174 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8" (OuterVolumeSpecName: "kube-api-access-s7mj8") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "kube-api-access-s7mj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.285846 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.307560 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.331480 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.349909 4727 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.349952 4727 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/613114a6-4186-4c9d-ba7c-f5d7029006ab-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.349968 4727 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.349982 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7mj8\" (UniqueName: \"kubernetes.io/projected/613114a6-4186-4c9d-ba7c-f5d7029006ab-kube-api-access-s7mj8\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.349995 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.350006 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.356666 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data" (OuterVolumeSpecName: "config-data") pod "613114a6-4186-4c9d-ba7c-f5d7029006ab" (UID: "613114a6-4186-4c9d-ba7c-f5d7029006ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.452047 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613114a6-4186-4c9d-ba7c-f5d7029006ab-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.898457 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"613114a6-4186-4c9d-ba7c-f5d7029006ab","Type":"ContainerDied","Data":"335a735f66295d2cf72cf0aae4b3f01a2c94d13a0ae09713263594a54e08c1c2"} Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.898523 4727 scope.go:117] "RemoveContainer" containerID="a5f59d1c429ff590de47c6d320f27917f0645433ee57cad83086b9161566c13d" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.898567 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.921999 4727 scope.go:117] "RemoveContainer" containerID="bc918bdbcb9b4f0ec39620940254dc61aed8d8a0c07d680c4b30c0bf0777bf3c" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.936493 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.952588 4727 scope.go:117] "RemoveContainer" containerID="f15f2a5e4ae8fca3078ed7192266eefe47a5bf68649866573a3761ddb8b558ae" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.955391 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.965464 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:58 crc kubenswrapper[4727]: E0929 10:43:58.965853 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="sg-core" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.965868 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="sg-core" Sep 29 10:43:58 crc kubenswrapper[4727]: E0929 10:43:58.965880 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-central-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.965885 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-central-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: E0929 10:43:58.965895 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="proxy-httpd" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.965902 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="proxy-httpd" Sep 29 10:43:58 crc kubenswrapper[4727]: E0929 10:43:58.965927 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-notification-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.965933 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-notification-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.966105 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="proxy-httpd" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.966124 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-notification-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.966140 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="sg-core" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.966150 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" containerName="ceilometer-central-agent" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.969612 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.971746 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.971826 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.973615 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.978407 4727 scope.go:117] "RemoveContainer" containerID="78bb6ff6d9c67306b0f27f5f7eacc5652e8ed2b48a081d56a949798e4a3112df" Sep 29 10:43:58 crc kubenswrapper[4727]: I0929 10:43:58.981581 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.063754 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2t9w\" (UniqueName: \"kubernetes.io/projected/aa89dc0c-297d-491c-a52d-41e14a330179-kube-api-access-p2t9w\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064102 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-log-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064156 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-config-data\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064183 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064284 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-scripts\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064315 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-run-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064369 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.064413 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.121901 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="613114a6-4186-4c9d-ba7c-f5d7029006ab" path="/var/lib/kubelet/pods/613114a6-4186-4c9d-ba7c-f5d7029006ab/volumes" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166022 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166140 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2t9w\" (UniqueName: \"kubernetes.io/projected/aa89dc0c-297d-491c-a52d-41e14a330179-kube-api-access-p2t9w\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166196 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-log-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166233 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-config-data\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166256 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166313 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-scripts\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166437 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-run-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166864 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166900 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-run-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.166903 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa89dc0c-297d-491c-a52d-41e14a330179-log-httpd\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.170392 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.170801 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.179438 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-scripts\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.180575 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-config-data\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.181778 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa89dc0c-297d-491c-a52d-41e14a330179-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.182883 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2t9w\" (UniqueName: \"kubernetes.io/projected/aa89dc0c-297d-491c-a52d-41e14a330179-kube-api-access-p2t9w\") pod \"ceilometer-0\" (UID: \"aa89dc0c-297d-491c-a52d-41e14a330179\") " pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.304950 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.739691 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.744946 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:43:59 crc kubenswrapper[4727]: I0929 10:43:59.907653 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa89dc0c-297d-491c-a52d-41e14a330179","Type":"ContainerStarted","Data":"ac7e39a6ab8e96209fa862552f6b916454cf8c14795fde74bcdfa9dc35db32c0"} Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.495649 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.593288 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data\") pod \"11290f51-f732-4756-8e1e-701a8699464f\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.593376 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m8p6\" (UniqueName: \"kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6\") pod \"11290f51-f732-4756-8e1e-701a8699464f\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.593516 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs\") pod \"11290f51-f732-4756-8e1e-701a8699464f\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.593662 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle\") pod \"11290f51-f732-4756-8e1e-701a8699464f\" (UID: \"11290f51-f732-4756-8e1e-701a8699464f\") " Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.594637 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs" (OuterVolumeSpecName: "logs") pod "11290f51-f732-4756-8e1e-701a8699464f" (UID: "11290f51-f732-4756-8e1e-701a8699464f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.599530 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6" (OuterVolumeSpecName: "kube-api-access-6m8p6") pod "11290f51-f732-4756-8e1e-701a8699464f" (UID: "11290f51-f732-4756-8e1e-701a8699464f"). InnerVolumeSpecName "kube-api-access-6m8p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.630119 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data" (OuterVolumeSpecName: "config-data") pod "11290f51-f732-4756-8e1e-701a8699464f" (UID: "11290f51-f732-4756-8e1e-701a8699464f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.639897 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11290f51-f732-4756-8e1e-701a8699464f" (UID: "11290f51-f732-4756-8e1e-701a8699464f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.697358 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11290f51-f732-4756-8e1e-701a8699464f-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.697390 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.697400 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11290f51-f732-4756-8e1e-701a8699464f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.697410 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m8p6\" (UniqueName: \"kubernetes.io/projected/11290f51-f732-4756-8e1e-701a8699464f-kube-api-access-6m8p6\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.922601 4727 generic.go:334] "Generic (PLEG): container finished" podID="11290f51-f732-4756-8e1e-701a8699464f" containerID="d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5" exitCode=0 Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.922684 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerDied","Data":"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5"} Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.922719 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"11290f51-f732-4756-8e1e-701a8699464f","Type":"ContainerDied","Data":"41071cbbad8659bc5e281e98c99303979f351a64b841029237b481d94ecca807"} Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.922740 4727 scope.go:117] "RemoveContainer" containerID="d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.922892 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.928060 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa89dc0c-297d-491c-a52d-41e14a330179","Type":"ContainerStarted","Data":"5633db7fba5a98af8700729fa6a645da6eafca7ed05d225e2ea7364d900efa35"} Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.965932 4727 scope.go:117] "RemoveContainer" containerID="970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.966949 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.982667 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996310 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996362 4727 scope.go:117] "RemoveContainer" containerID="d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5" Sep 29 10:44:00 crc kubenswrapper[4727]: E0929 10:44:00.996763 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5\": container with ID starting with d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5 not found: ID does not exist" containerID="d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5" Sep 29 10:44:00 crc kubenswrapper[4727]: E0929 10:44:00.996788 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-api" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996803 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-api" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996804 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5"} err="failed to get container status \"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5\": rpc error: code = NotFound desc = could not find container \"d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5\": container with ID starting with d0f0d7b962eb477967236d48a66d9a26cb898c6bee45f8177eb42ad565ad39f5 not found: ID does not exist" Sep 29 10:44:00 crc kubenswrapper[4727]: E0929 10:44:00.996842 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-log" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996851 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-log" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.996865 4727 scope.go:117] "RemoveContainer" containerID="970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.997013 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-log" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.997045 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="11290f51-f732-4756-8e1e-701a8699464f" containerName="nova-api-api" Sep 29 10:44:00 crc kubenswrapper[4727]: E0929 10:44:00.997295 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4\": container with ID starting with 970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4 not found: ID does not exist" containerID="970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.997315 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4"} err="failed to get container status \"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4\": rpc error: code = NotFound desc = could not find container \"970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4\": container with ID starting with 970f4764b4d3714e26d8d002f2950aaa9d4d36ad4332f049e409eb51d8201de4 not found: ID does not exist" Sep 29 10:44:00 crc kubenswrapper[4727]: I0929 10:44:00.998299 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.001054 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.001759 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.004692 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.009598 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106084 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106238 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106268 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106298 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106326 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.106355 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxqzx\" (UniqueName: \"kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.123959 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11290f51-f732-4756-8e1e-701a8699464f" path="/var/lib/kubelet/pods/11290f51-f732-4756-8e1e-701a8699464f/volumes" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.207605 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.207985 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.208012 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.208045 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.208062 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.208076 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxqzx\" (UniqueName: \"kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.208275 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.212351 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.212788 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.213056 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.229364 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxqzx\" (UniqueName: \"kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.232482 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs\") pod \"nova-api-0\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.247526 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.271490 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.330777 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.810010 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:01 crc kubenswrapper[4727]: W0929 10:44:01.816711 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod003d340e_c930_4e63_b094_f46ff3b92f67.slice/crio-8b721e700293b9d6fb40e0766816783f71260d4116817969c9d80b5c1f2ba2b5 WatchSource:0}: Error finding container 8b721e700293b9d6fb40e0766816783f71260d4116817969c9d80b5c1f2ba2b5: Status 404 returned error can't find the container with id 8b721e700293b9d6fb40e0766816783f71260d4116817969c9d80b5c1f2ba2b5 Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.940020 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa89dc0c-297d-491c-a52d-41e14a330179","Type":"ContainerStarted","Data":"8a25d9db7ccb1257c275b217d27c9090186450766543f85569f718a53b5f7ff7"} Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.941697 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerStarted","Data":"8b721e700293b9d6fb40e0766816783f71260d4116817969c9d80b5c1f2ba2b5"} Sep 29 10:44:01 crc kubenswrapper[4727]: I0929 10:44:01.963919 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.151225 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-m77vs"] Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.152972 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.156835 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.156849 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.163668 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m77vs"] Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.229699 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9c5\" (UniqueName: \"kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.229761 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.229833 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.229923 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.331514 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.331909 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf9c5\" (UniqueName: \"kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.331946 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.331989 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.337940 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.338119 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.346786 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.350777 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf9c5\" (UniqueName: \"kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5\") pod \"nova-cell1-cell-mapping-m77vs\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.508188 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.951414 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa89dc0c-297d-491c-a52d-41e14a330179","Type":"ContainerStarted","Data":"4bbdc163061b82603886fce5d77990d14b1317376b0f54ece00431df44450aaa"} Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.953219 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerStarted","Data":"ae0eafef6a710639b4b065e74df87077d6191590b90be7ac80b73a196f5ed012"} Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.953462 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerStarted","Data":"4c4ac76a748ddd35e4427c965427b15abf573ba62529c8c2eed995c11b2d6eab"} Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.974868 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.974842192 podStartE2EDuration="2.974842192s" podCreationTimestamp="2025-09-29 10:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:44:02.969159675 +0000 UTC m=+1313.142473067" watchObservedRunningTime="2025-09-29 10:44:02.974842192 +0000 UTC m=+1313.148155554" Sep 29 10:44:02 crc kubenswrapper[4727]: I0929 10:44:02.999205 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m77vs"] Sep 29 10:44:03 crc kubenswrapper[4727]: W0929 10:44:03.008634 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2 WatchSource:0}: Error finding container 1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2: Status 404 returned error can't find the container with id 1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2 Sep 29 10:44:03 crc kubenswrapper[4727]: I0929 10:44:03.967428 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa89dc0c-297d-491c-a52d-41e14a330179","Type":"ContainerStarted","Data":"5a0950823f86da39fc63e5f4a2e805f82689d05f8af301d875c650fb4ca97b89"} Sep 29 10:44:03 crc kubenswrapper[4727]: I0929 10:44:03.968025 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 29 10:44:03 crc kubenswrapper[4727]: I0929 10:44:03.971347 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m77vs" event={"ID":"5cbd9580-fa48-48d9-bcf1-4009c21d31a2","Type":"ContainerStarted","Data":"8003c89863fa6c45ea034d993d33632f1c50aeb576150ec5a450e490934ebcff"} Sep 29 10:44:03 crc kubenswrapper[4727]: I0929 10:44:03.971385 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m77vs" event={"ID":"5cbd9580-fa48-48d9-bcf1-4009c21d31a2","Type":"ContainerStarted","Data":"1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2"} Sep 29 10:44:03 crc kubenswrapper[4727]: I0929 10:44:03.989874 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.624965007 podStartE2EDuration="5.989856982s" podCreationTimestamp="2025-09-29 10:43:58 +0000 UTC" firstStartedPulling="2025-09-29 10:43:59.744701721 +0000 UTC m=+1309.918015083" lastFinishedPulling="2025-09-29 10:44:03.109593696 +0000 UTC m=+1313.282907058" observedRunningTime="2025-09-29 10:44:03.987252495 +0000 UTC m=+1314.160565857" watchObservedRunningTime="2025-09-29 10:44:03.989856982 +0000 UTC m=+1314.163170344" Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.004503 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-m77vs" podStartSLOduration=2.004482819 podStartE2EDuration="2.004482819s" podCreationTimestamp="2025-09-29 10:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:44:04.002355994 +0000 UTC m=+1314.175669356" watchObservedRunningTime="2025-09-29 10:44:04.004482819 +0000 UTC m=+1314.177796181" Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.422524 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.532262 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.532606 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="dnsmasq-dns" containerID="cri-o://aee4fd018b4ccd735a6280897bf356facb5eb8006522a7d66a8b7d462d741804" gracePeriod=10 Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.988304 4727 generic.go:334] "Generic (PLEG): container finished" podID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerID="aee4fd018b4ccd735a6280897bf356facb5eb8006522a7d66a8b7d462d741804" exitCode=0 Sep 29 10:44:04 crc kubenswrapper[4727]: I0929 10:44:04.988371 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerDied","Data":"aee4fd018b4ccd735a6280897bf356facb5eb8006522a7d66a8b7d462d741804"} Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.084549 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.193:5353: connect: connection refused" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.482671 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.598997 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmhzc\" (UniqueName: \"kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.599131 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.599153 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.599193 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.599305 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.599361 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc\") pod \"dd841b62-c07f-47bc-8da0-1f36a1e51176\" (UID: \"dd841b62-c07f-47bc-8da0-1f36a1e51176\") " Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.607785 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc" (OuterVolumeSpecName: "kube-api-access-gmhzc") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "kube-api-access-gmhzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.682755 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.685821 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.695990 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.702697 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.702740 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.702754 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.702768 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmhzc\" (UniqueName: \"kubernetes.io/projected/dd841b62-c07f-47bc-8da0-1f36a1e51176-kube-api-access-gmhzc\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.704852 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.714906 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config" (OuterVolumeSpecName: "config") pod "dd841b62-c07f-47bc-8da0-1f36a1e51176" (UID: "dd841b62-c07f-47bc-8da0-1f36a1e51176"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.804197 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:05 crc kubenswrapper[4727]: I0929 10:44:05.804516 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd841b62-c07f-47bc-8da0-1f36a1e51176-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.003951 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" event={"ID":"dd841b62-c07f-47bc-8da0-1f36a1e51176","Type":"ContainerDied","Data":"67774f38e7b0d0a8ae9fa72d82d3dd51a1502b17c8f363a2cda0d5c2a0f9cae3"} Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.003999 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699d7ddff-6tqz6" Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.004018 4727 scope.go:117] "RemoveContainer" containerID="aee4fd018b4ccd735a6280897bf356facb5eb8006522a7d66a8b7d462d741804" Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.033286 4727 scope.go:117] "RemoveContainer" containerID="407346dd0b039458a59e185d34fe7429e16bca7aac4b43154d9c47b35e321f67" Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.067190 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:44:06 crc kubenswrapper[4727]: I0929 10:44:06.073868 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-699d7ddff-6tqz6"] Sep 29 10:44:07 crc kubenswrapper[4727]: I0929 10:44:07.120568 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" path="/var/lib/kubelet/pods/dd841b62-c07f-47bc-8da0-1f36a1e51176/volumes" Sep 29 10:44:09 crc kubenswrapper[4727]: I0929 10:44:09.077358 4727 generic.go:334] "Generic (PLEG): container finished" podID="5cbd9580-fa48-48d9-bcf1-4009c21d31a2" containerID="8003c89863fa6c45ea034d993d33632f1c50aeb576150ec5a450e490934ebcff" exitCode=0 Sep 29 10:44:09 crc kubenswrapper[4727]: I0929 10:44:09.077464 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m77vs" event={"ID":"5cbd9580-fa48-48d9-bcf1-4009c21d31a2","Type":"ContainerDied","Data":"8003c89863fa6c45ea034d993d33632f1c50aeb576150ec5a450e490934ebcff"} Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.457127 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.510305 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle\") pod \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.510402 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf9c5\" (UniqueName: \"kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5\") pod \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.510608 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data\") pod \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.510699 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts\") pod \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\" (UID: \"5cbd9580-fa48-48d9-bcf1-4009c21d31a2\") " Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.516789 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5" (OuterVolumeSpecName: "kube-api-access-hf9c5") pod "5cbd9580-fa48-48d9-bcf1-4009c21d31a2" (UID: "5cbd9580-fa48-48d9-bcf1-4009c21d31a2"). InnerVolumeSpecName "kube-api-access-hf9c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.519527 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts" (OuterVolumeSpecName: "scripts") pod "5cbd9580-fa48-48d9-bcf1-4009c21d31a2" (UID: "5cbd9580-fa48-48d9-bcf1-4009c21d31a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.537765 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data" (OuterVolumeSpecName: "config-data") pod "5cbd9580-fa48-48d9-bcf1-4009c21d31a2" (UID: "5cbd9580-fa48-48d9-bcf1-4009c21d31a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.537711 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cbd9580-fa48-48d9-bcf1-4009c21d31a2" (UID: "5cbd9580-fa48-48d9-bcf1-4009c21d31a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.614032 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.614076 4727 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-scripts\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.614090 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:10 crc kubenswrapper[4727]: I0929 10:44:10.614107 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf9c5\" (UniqueName: \"kubernetes.io/projected/5cbd9580-fa48-48d9-bcf1-4009c21d31a2-kube-api-access-hf9c5\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.099792 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m77vs" event={"ID":"5cbd9580-fa48-48d9-bcf1-4009c21d31a2","Type":"ContainerDied","Data":"1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2"} Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.099853 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2" Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.099896 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m77vs" Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.281634 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.281913 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-log" containerID="cri-o://4c4ac76a748ddd35e4427c965427b15abf573ba62529c8c2eed995c11b2d6eab" gracePeriod=30 Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.282079 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-api" containerID="cri-o://ae0eafef6a710639b4b065e74df87077d6191590b90be7ac80b73a196f5ed012" gracePeriod=30 Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.292703 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.293156 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a0db3ea4-126f-4c58-bbd4-859916d5347f" containerName="nova-scheduler-scheduler" containerID="cri-o://3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243" gracePeriod=30 Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.340508 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.340773 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" containerID="cri-o://e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9" gracePeriod=30 Sep 29 10:44:11 crc kubenswrapper[4727]: I0929 10:44:11.340931 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" containerID="cri-o://fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2" gracePeriod=30 Sep 29 10:44:11 crc kubenswrapper[4727]: E0929 10:44:11.742563 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache]" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.126157 4727 generic.go:334] "Generic (PLEG): container finished" podID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerID="e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9" exitCode=143 Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.126234 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerDied","Data":"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9"} Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.129163 4727 generic.go:334] "Generic (PLEG): container finished" podID="003d340e-c930-4e63-b094-f46ff3b92f67" containerID="ae0eafef6a710639b4b065e74df87077d6191590b90be7ac80b73a196f5ed012" exitCode=0 Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.129197 4727 generic.go:334] "Generic (PLEG): container finished" podID="003d340e-c930-4e63-b094-f46ff3b92f67" containerID="4c4ac76a748ddd35e4427c965427b15abf573ba62529c8c2eed995c11b2d6eab" exitCode=143 Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.129220 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerDied","Data":"ae0eafef6a710639b4b065e74df87077d6191590b90be7ac80b73a196f5ed012"} Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.129248 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerDied","Data":"4c4ac76a748ddd35e4427c965427b15abf573ba62529c8c2eed995c11b2d6eab"} Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.436783 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.559691 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.560545 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs" (OuterVolumeSpecName: "logs") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.560717 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxqzx\" (UniqueName: \"kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.560753 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.561535 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.561653 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.561711 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs\") pod \"003d340e-c930-4e63-b094-f46ff3b92f67\" (UID: \"003d340e-c930-4e63-b094-f46ff3b92f67\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.562482 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003d340e-c930-4e63-b094-f46ff3b92f67-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.567811 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx" (OuterVolumeSpecName: "kube-api-access-kxqzx") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "kube-api-access-kxqzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.587572 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.591069 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data" (OuterVolumeSpecName: "config-data") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.612657 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.614515 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "003d340e-c930-4e63-b094-f46ff3b92f67" (UID: "003d340e-c930-4e63-b094-f46ff3b92f67"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.658850 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.663917 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.663946 4727 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.663960 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxqzx\" (UniqueName: \"kubernetes.io/projected/003d340e-c930-4e63-b094-f46ff3b92f67-kube-api-access-kxqzx\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.663973 4727 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.663984 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d340e-c930-4e63-b094-f46ff3b92f67-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.765284 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data\") pod \"a0db3ea4-126f-4c58-bbd4-859916d5347f\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.765497 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn9sj\" (UniqueName: \"kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj\") pod \"a0db3ea4-126f-4c58-bbd4-859916d5347f\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.765638 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle\") pod \"a0db3ea4-126f-4c58-bbd4-859916d5347f\" (UID: \"a0db3ea4-126f-4c58-bbd4-859916d5347f\") " Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.771660 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj" (OuterVolumeSpecName: "kube-api-access-rn9sj") pod "a0db3ea4-126f-4c58-bbd4-859916d5347f" (UID: "a0db3ea4-126f-4c58-bbd4-859916d5347f"). InnerVolumeSpecName "kube-api-access-rn9sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.790555 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data" (OuterVolumeSpecName: "config-data") pod "a0db3ea4-126f-4c58-bbd4-859916d5347f" (UID: "a0db3ea4-126f-4c58-bbd4-859916d5347f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.791915 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0db3ea4-126f-4c58-bbd4-859916d5347f" (UID: "a0db3ea4-126f-4c58-bbd4-859916d5347f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.868209 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn9sj\" (UniqueName: \"kubernetes.io/projected/a0db3ea4-126f-4c58-bbd4-859916d5347f-kube-api-access-rn9sj\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.868248 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:12 crc kubenswrapper[4727]: I0929 10:44:12.868261 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0db3ea4-126f-4c58-bbd4-859916d5347f-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.149026 4727 generic.go:334] "Generic (PLEG): container finished" podID="a0db3ea4-126f-4c58-bbd4-859916d5347f" containerID="3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243" exitCode=0 Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.149085 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0db3ea4-126f-4c58-bbd4-859916d5347f","Type":"ContainerDied","Data":"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243"} Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.149111 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0db3ea4-126f-4c58-bbd4-859916d5347f","Type":"ContainerDied","Data":"4985f1c4ce491b10e43282430247dda2259970f4bf65736732aa17bd44dfb264"} Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.149126 4727 scope.go:117] "RemoveContainer" containerID="3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.149232 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.157041 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"003d340e-c930-4e63-b094-f46ff3b92f67","Type":"ContainerDied","Data":"8b721e700293b9d6fb40e0766816783f71260d4116817969c9d80b5c1f2ba2b5"} Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.157133 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.173266 4727 scope.go:117] "RemoveContainer" containerID="3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.177403 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243\": container with ID starting with 3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243 not found: ID does not exist" containerID="3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.177462 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243"} err="failed to get container status \"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243\": rpc error: code = NotFound desc = could not find container \"3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243\": container with ID starting with 3674e49e847106ba8675a288a0eca6ccc6c04ffb7646bfaa5abb7e3c69414243 not found: ID does not exist" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.177493 4727 scope.go:117] "RemoveContainer" containerID="ae0eafef6a710639b4b065e74df87077d6191590b90be7ac80b73a196f5ed012" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.184548 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.198056 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.208684 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.224824 4727 scope.go:117] "RemoveContainer" containerID="4c4ac76a748ddd35e4427c965427b15abf573ba62529c8c2eed995c11b2d6eab" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.225405 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.234278 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.234868 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0db3ea4-126f-4c58-bbd4-859916d5347f" containerName="nova-scheduler-scheduler" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.234880 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0db3ea4-126f-4c58-bbd4-859916d5347f" containerName="nova-scheduler-scheduler" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.234919 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-api" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.234925 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-api" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.234932 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbd9580-fa48-48d9-bcf1-4009c21d31a2" containerName="nova-manage" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.234939 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbd9580-fa48-48d9-bcf1-4009c21d31a2" containerName="nova-manage" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.234958 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="dnsmasq-dns" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.234963 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="dnsmasq-dns" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.235014 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-log" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235021 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-log" Sep 29 10:44:13 crc kubenswrapper[4727]: E0929 10:44:13.235031 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="init" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235037 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="init" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235269 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-log" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235286 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbd9580-fa48-48d9-bcf1-4009c21d31a2" containerName="nova-manage" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235304 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0db3ea4-126f-4c58-bbd4-859916d5347f" containerName="nova-scheduler-scheduler" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235316 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd841b62-c07f-47bc-8da0-1f36a1e51176" containerName="dnsmasq-dns" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.235346 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" containerName="nova-api-api" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.236029 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.237676 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.246844 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.255859 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.257633 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.261230 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.261814 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.261996 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.272554 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381357 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-config-data\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381799 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381856 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381897 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-config-data\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381922 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/725e53c7-0270-451e-8bde-851f9a60a8ab-logs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.381973 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7pl4\" (UniqueName: \"kubernetes.io/projected/41ff8f24-0d46-49e0-88be-7b4d3c33d619-kube-api-access-p7pl4\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.382040 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.382097 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.382146 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs6m4\" (UniqueName: \"kubernetes.io/projected/725e53c7-0270-451e-8bde-851f9a60a8ab-kube-api-access-rs6m4\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483292 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs6m4\" (UniqueName: \"kubernetes.io/projected/725e53c7-0270-451e-8bde-851f9a60a8ab-kube-api-access-rs6m4\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483491 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-config-data\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483542 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483568 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483594 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-config-data\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483611 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/725e53c7-0270-451e-8bde-851f9a60a8ab-logs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483651 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7pl4\" (UniqueName: \"kubernetes.io/projected/41ff8f24-0d46-49e0-88be-7b4d3c33d619-kube-api-access-p7pl4\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483688 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.483729 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.484361 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/725e53c7-0270-451e-8bde-851f9a60a8ab-logs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.489727 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-config-data\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.489786 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.489927 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.490216 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-config-data\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.490515 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ff8f24-0d46-49e0-88be-7b4d3c33d619-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.492901 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/725e53c7-0270-451e-8bde-851f9a60a8ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.503100 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs6m4\" (UniqueName: \"kubernetes.io/projected/725e53c7-0270-451e-8bde-851f9a60a8ab-kube-api-access-rs6m4\") pod \"nova-api-0\" (UID: \"725e53c7-0270-451e-8bde-851f9a60a8ab\") " pod="openstack/nova-api-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.503574 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7pl4\" (UniqueName: \"kubernetes.io/projected/41ff8f24-0d46-49e0-88be-7b4d3c33d619-kube-api-access-p7pl4\") pod \"nova-scheduler-0\" (UID: \"41ff8f24-0d46-49e0-88be-7b4d3c33d619\") " pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.595404 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 29 10:44:13 crc kubenswrapper[4727]: I0929 10:44:13.611721 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.082563 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 29 10:44:14 crc kubenswrapper[4727]: W0929 10:44:14.088765 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41ff8f24_0d46_49e0_88be_7b4d3c33d619.slice/crio-15a84afd53fccb4ab2a60659edd88c33c8556a055857522de745ceefe2f978e9 WatchSource:0}: Error finding container 15a84afd53fccb4ab2a60659edd88c33c8556a055857522de745ceefe2f978e9: Status 404 returned error can't find the container with id 15a84afd53fccb4ab2a60659edd88c33c8556a055857522de745ceefe2f978e9 Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.168465 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.168945 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41ff8f24-0d46-49e0-88be-7b4d3c33d619","Type":"ContainerStarted","Data":"15a84afd53fccb4ab2a60659edd88c33c8556a055857522de745ceefe2f978e9"} Sep 29 10:44:14 crc kubenswrapper[4727]: W0929 10:44:14.176534 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod725e53c7_0270_451e_8bde_851f9a60a8ab.slice/crio-868bbd4b38a3bd60f7ce77a3fe8ae73491ec0eaa024097e73a2bc6af5d2644ad WatchSource:0}: Error finding container 868bbd4b38a3bd60f7ce77a3fe8ae73491ec0eaa024097e73a2bc6af5d2644ad: Status 404 returned error can't find the container with id 868bbd4b38a3bd60f7ce77a3fe8ae73491ec0eaa024097e73a2bc6af5d2644ad Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.470735 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:40642->10.217.0.198:8775: read: connection reset by peer" Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.470814 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:40644->10.217.0.198:8775: read: connection reset by peer" Sep 29 10:44:14 crc kubenswrapper[4727]: I0929 10:44:14.894087 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.015385 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs\") pod \"2c2a19c0-67ab-4406-8a75-e18012681d9d\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.015510 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data\") pod \"2c2a19c0-67ab-4406-8a75-e18012681d9d\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.015532 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs\") pod \"2c2a19c0-67ab-4406-8a75-e18012681d9d\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.015581 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98mls\" (UniqueName: \"kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls\") pod \"2c2a19c0-67ab-4406-8a75-e18012681d9d\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.015621 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle\") pod \"2c2a19c0-67ab-4406-8a75-e18012681d9d\" (UID: \"2c2a19c0-67ab-4406-8a75-e18012681d9d\") " Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.017319 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs" (OuterVolumeSpecName: "logs") pod "2c2a19c0-67ab-4406-8a75-e18012681d9d" (UID: "2c2a19c0-67ab-4406-8a75-e18012681d9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.026651 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls" (OuterVolumeSpecName: "kube-api-access-98mls") pod "2c2a19c0-67ab-4406-8a75-e18012681d9d" (UID: "2c2a19c0-67ab-4406-8a75-e18012681d9d"). InnerVolumeSpecName "kube-api-access-98mls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.047835 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data" (OuterVolumeSpecName: "config-data") pod "2c2a19c0-67ab-4406-8a75-e18012681d9d" (UID: "2c2a19c0-67ab-4406-8a75-e18012681d9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.050561 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c2a19c0-67ab-4406-8a75-e18012681d9d" (UID: "2c2a19c0-67ab-4406-8a75-e18012681d9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.100327 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2c2a19c0-67ab-4406-8a75-e18012681d9d" (UID: "2c2a19c0-67ab-4406-8a75-e18012681d9d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.118079 4727 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c2a19c0-67ab-4406-8a75-e18012681d9d-logs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.119999 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.120088 4727 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.120161 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98mls\" (UniqueName: \"kubernetes.io/projected/2c2a19c0-67ab-4406-8a75-e18012681d9d-kube-api-access-98mls\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.120221 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c2a19c0-67ab-4406-8a75-e18012681d9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.123542 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003d340e-c930-4e63-b094-f46ff3b92f67" path="/var/lib/kubelet/pods/003d340e-c930-4e63-b094-f46ff3b92f67/volumes" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.124712 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0db3ea4-126f-4c58-bbd4-859916d5347f" path="/var/lib/kubelet/pods/a0db3ea4-126f-4c58-bbd4-859916d5347f/volumes" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.178449 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41ff8f24-0d46-49e0-88be-7b4d3c33d619","Type":"ContainerStarted","Data":"2f379bf79f01c57990cf904cf790dceb6ddda82c71f452212e07bea3e854e205"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.180520 4727 generic.go:334] "Generic (PLEG): container finished" podID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerID="fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2" exitCode=0 Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.180572 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.180594 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerDied","Data":"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.180621 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c2a19c0-67ab-4406-8a75-e18012681d9d","Type":"ContainerDied","Data":"d1835b28eb65e2df607f8fa2e3ef1453cc0c9640f4e4104f3c685d9e82f9b0ce"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.180641 4727 scope.go:117] "RemoveContainer" containerID="fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.184448 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"725e53c7-0270-451e-8bde-851f9a60a8ab","Type":"ContainerStarted","Data":"d48c0b87cabcac4efea7a47f9f80386d05a2b6cac290ddcd3b58133adf96def6"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.184487 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"725e53c7-0270-451e-8bde-851f9a60a8ab","Type":"ContainerStarted","Data":"4e32e960ecb7d010d630a99a794563615dd990763aaf1c956769b7446ae20448"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.184498 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"725e53c7-0270-451e-8bde-851f9a60a8ab","Type":"ContainerStarted","Data":"868bbd4b38a3bd60f7ce77a3fe8ae73491ec0eaa024097e73a2bc6af5d2644ad"} Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.194438 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.194417733 podStartE2EDuration="2.194417733s" podCreationTimestamp="2025-09-29 10:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:44:15.192163945 +0000 UTC m=+1325.365477327" watchObservedRunningTime="2025-09-29 10:44:15.194417733 +0000 UTC m=+1325.367731095" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.224795 4727 scope.go:117] "RemoveContainer" containerID="e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.242435 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.251751 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.261274 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2612583969999998 podStartE2EDuration="2.261258397s" podCreationTimestamp="2025-09-29 10:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:44:15.21911777 +0000 UTC m=+1325.392431162" watchObservedRunningTime="2025-09-29 10:44:15.261258397 +0000 UTC m=+1325.434571759" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.263053 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:15 crc kubenswrapper[4727]: E0929 10:44:15.263571 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.263587 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" Sep 29 10:44:15 crc kubenswrapper[4727]: E0929 10:44:15.263608 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.263617 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.263888 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-metadata" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.263916 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" containerName="nova-metadata-log" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.265166 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.268567 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.268741 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.293479 4727 scope.go:117] "RemoveContainer" containerID="fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2" Sep 29 10:44:15 crc kubenswrapper[4727]: E0929 10:44:15.294105 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2\": container with ID starting with fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2 not found: ID does not exist" containerID="fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.294178 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2"} err="failed to get container status \"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2\": rpc error: code = NotFound desc = could not find container \"fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2\": container with ID starting with fac803198d26a40333e681993bc1d10e652164c689ec6b2e7a3fba78d2883ab2 not found: ID does not exist" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.294218 4727 scope.go:117] "RemoveContainer" containerID="e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9" Sep 29 10:44:15 crc kubenswrapper[4727]: E0929 10:44:15.294581 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9\": container with ID starting with e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9 not found: ID does not exist" containerID="e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.294648 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9"} err="failed to get container status \"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9\": rpc error: code = NotFound desc = could not find container \"e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9\": container with ID starting with e0edbee86afdd762c88e89f39f8ce60f1bb87960e3fc76a8704e89cfd5cc83b9 not found: ID does not exist" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.307162 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.328155 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f944aa92-5fe8-4100-9faf-00ef0443d80f-logs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.328270 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.328368 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz8ll\" (UniqueName: \"kubernetes.io/projected/f944aa92-5fe8-4100-9faf-00ef0443d80f-kube-api-access-mz8ll\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.328555 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-config-data\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.328629 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.430972 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f944aa92-5fe8-4100-9faf-00ef0443d80f-logs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.431061 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.431105 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz8ll\" (UniqueName: \"kubernetes.io/projected/f944aa92-5fe8-4100-9faf-00ef0443d80f-kube-api-access-mz8ll\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.431505 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-config-data\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.431554 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.431608 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f944aa92-5fe8-4100-9faf-00ef0443d80f-logs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.435611 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.436491 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-config-data\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.437763 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f944aa92-5fe8-4100-9faf-00ef0443d80f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.448591 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz8ll\" (UniqueName: \"kubernetes.io/projected/f944aa92-5fe8-4100-9faf-00ef0443d80f-kube-api-access-mz8ll\") pod \"nova-metadata-0\" (UID: \"f944aa92-5fe8-4100-9faf-00ef0443d80f\") " pod="openstack/nova-metadata-0" Sep 29 10:44:15 crc kubenswrapper[4727]: I0929 10:44:15.592847 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 29 10:44:16 crc kubenswrapper[4727]: W0929 10:44:16.036613 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf944aa92_5fe8_4100_9faf_00ef0443d80f.slice/crio-64c82dc2bdbd9c947968cf4e0d79e7654ce7bccdb5fe1af8a676021adc232b10 WatchSource:0}: Error finding container 64c82dc2bdbd9c947968cf4e0d79e7654ce7bccdb5fe1af8a676021adc232b10: Status 404 returned error can't find the container with id 64c82dc2bdbd9c947968cf4e0d79e7654ce7bccdb5fe1af8a676021adc232b10 Sep 29 10:44:16 crc kubenswrapper[4727]: I0929 10:44:16.039953 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 29 10:44:16 crc kubenswrapper[4727]: I0929 10:44:16.194436 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f944aa92-5fe8-4100-9faf-00ef0443d80f","Type":"ContainerStarted","Data":"64c82dc2bdbd9c947968cf4e0d79e7654ce7bccdb5fe1af8a676021adc232b10"} Sep 29 10:44:17 crc kubenswrapper[4727]: I0929 10:44:17.120722 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c2a19c0-67ab-4406-8a75-e18012681d9d" path="/var/lib/kubelet/pods/2c2a19c0-67ab-4406-8a75-e18012681d9d/volumes" Sep 29 10:44:17 crc kubenswrapper[4727]: I0929 10:44:17.211815 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f944aa92-5fe8-4100-9faf-00ef0443d80f","Type":"ContainerStarted","Data":"4d75b61796b6b4a412bf7438bb9108e9bae2cfd5da87c48c0eb2640f2c33b0de"} Sep 29 10:44:17 crc kubenswrapper[4727]: I0929 10:44:17.211892 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f944aa92-5fe8-4100-9faf-00ef0443d80f","Type":"ContainerStarted","Data":"aeb3d26a350bacb7e99807b0d4c9ede36795cc87e8bef639fa208e94e5f8d2a8"} Sep 29 10:44:17 crc kubenswrapper[4727]: I0929 10:44:17.243093 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.243066983 podStartE2EDuration="2.243066983s" podCreationTimestamp="2025-09-29 10:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:44:17.234600324 +0000 UTC m=+1327.407913756" watchObservedRunningTime="2025-09-29 10:44:17.243066983 +0000 UTC m=+1327.416380355" Sep 29 10:44:18 crc kubenswrapper[4727]: I0929 10:44:18.595687 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 29 10:44:19 crc kubenswrapper[4727]: I0929 10:44:19.246331 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:44:19 crc kubenswrapper[4727]: I0929 10:44:19.246666 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:44:20 crc kubenswrapper[4727]: I0929 10:44:20.593506 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:44:20 crc kubenswrapper[4727]: I0929 10:44:20.593604 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 29 10:44:22 crc kubenswrapper[4727]: E0929 10:44:22.003558 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache]" Sep 29 10:44:23 crc kubenswrapper[4727]: I0929 10:44:23.595961 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 29 10:44:23 crc kubenswrapper[4727]: I0929 10:44:23.611903 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:44:23 crc kubenswrapper[4727]: I0929 10:44:23.611991 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 29 10:44:23 crc kubenswrapper[4727]: I0929 10:44:23.632632 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 29 10:44:24 crc kubenswrapper[4727]: I0929 10:44:24.354928 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 29 10:44:24 crc kubenswrapper[4727]: I0929 10:44:24.628470 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="725e53c7-0270-451e-8bde-851f9a60a8ab" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 10:44:24 crc kubenswrapper[4727]: I0929 10:44:24.628470 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="725e53c7-0270-451e-8bde-851f9a60a8ab" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 10:44:25 crc kubenswrapper[4727]: I0929 10:44:25.593865 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 10:44:25 crc kubenswrapper[4727]: I0929 10:44:25.594183 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 29 10:44:26 crc kubenswrapper[4727]: I0929 10:44:26.611466 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f944aa92-5fe8-4100-9faf-00ef0443d80f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 10:44:26 crc kubenswrapper[4727]: I0929 10:44:26.611474 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f944aa92-5fe8-4100-9faf-00ef0443d80f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 29 10:44:29 crc kubenswrapper[4727]: I0929 10:44:29.313557 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 29 10:44:32 crc kubenswrapper[4727]: E0929 10:44:32.292198 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache]" Sep 29 10:44:33 crc kubenswrapper[4727]: I0929 10:44:33.618141 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 10:44:33 crc kubenswrapper[4727]: I0929 10:44:33.619240 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 10:44:33 crc kubenswrapper[4727]: I0929 10:44:33.620591 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 29 10:44:33 crc kubenswrapper[4727]: I0929 10:44:33.625433 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 10:44:34 crc kubenswrapper[4727]: I0929 10:44:34.388815 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 29 10:44:34 crc kubenswrapper[4727]: I0929 10:44:34.394975 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 29 10:44:35 crc kubenswrapper[4727]: I0929 10:44:35.600535 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 10:44:35 crc kubenswrapper[4727]: I0929 10:44:35.602430 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 29 10:44:35 crc kubenswrapper[4727]: I0929 10:44:35.614737 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 10:44:36 crc kubenswrapper[4727]: I0929 10:44:36.413478 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 29 10:44:42 crc kubenswrapper[4727]: E0929 10:44:42.545422 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache]" Sep 29 10:44:44 crc kubenswrapper[4727]: I0929 10:44:44.385405 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:45 crc kubenswrapper[4727]: I0929 10:44:45.253038 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:48 crc kubenswrapper[4727]: I0929 10:44:48.559720 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="rabbitmq" containerID="cri-o://6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7" gracePeriod=604796 Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.247139 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.247207 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.247251 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.247992 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.248054 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315" gracePeriod=600 Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.335476 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="rabbitmq" containerID="cri-o://e3f4c3abce76e3e7604247f7d881163b52102d07e2ccfa590484a9575c282bf0" gracePeriod=604796 Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.532947 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315" exitCode=0 Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.532993 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315"} Sep 29 10:44:49 crc kubenswrapper[4727]: I0929 10:44:49.533027 4727 scope.go:117] "RemoveContainer" containerID="d1fdb01774e7fff6b0e920a0dae44896b281ab1e9c0fc1df615fecf2e9b9129b" Sep 29 10:44:50 crc kubenswrapper[4727]: I0929 10:44:50.542577 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4"} Sep 29 10:44:50 crc kubenswrapper[4727]: I0929 10:44:50.900778 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Sep 29 10:44:51 crc kubenswrapper[4727]: I0929 10:44:51.265329 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Sep 29 10:44:52 crc kubenswrapper[4727]: E0929 10:44:52.808083 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache]" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.126755 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.289754 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.289832 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.289913 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.289986 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290058 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290085 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pl5l\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290144 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290176 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290224 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290254 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290279 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie\") pod \"667cfa10-bc94-4788-af5d-296745385383\" (UID: \"667cfa10-bc94-4788-af5d-296745385383\") " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290640 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.290753 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.291179 4727 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.291200 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.291772 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.296605 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info" (OuterVolumeSpecName: "pod-info") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.296626 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.296656 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.297428 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l" (OuterVolumeSpecName: "kube-api-access-4pl5l") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "kube-api-access-4pl5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.299290 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.333686 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data" (OuterVolumeSpecName: "config-data") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.385946 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf" (OuterVolumeSpecName: "server-conf") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.392963 4727 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/667cfa10-bc94-4788-af5d-296745385383-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393013 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pl5l\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-kube-api-access-4pl5l\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393041 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393071 4727 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393079 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393090 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393097 4727 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/667cfa10-bc94-4788-af5d-296745385383-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.393107 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/667cfa10-bc94-4788-af5d-296745385383-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.414240 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.431634 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "667cfa10-bc94-4788-af5d-296745385383" (UID: "667cfa10-bc94-4788-af5d-296745385383"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.494391 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.494435 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/667cfa10-bc94-4788-af5d-296745385383-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.601266 4727 generic.go:334] "Generic (PLEG): container finished" podID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerID="e3f4c3abce76e3e7604247f7d881163b52102d07e2ccfa590484a9575c282bf0" exitCode=0 Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.601368 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerDied","Data":"e3f4c3abce76e3e7604247f7d881163b52102d07e2ccfa590484a9575c282bf0"} Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.604947 4727 generic.go:334] "Generic (PLEG): container finished" podID="667cfa10-bc94-4788-af5d-296745385383" containerID="6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7" exitCode=0 Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.605005 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerDied","Data":"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7"} Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.605041 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"667cfa10-bc94-4788-af5d-296745385383","Type":"ContainerDied","Data":"b910c6dbfec64243cf3bf137a89f4a36681499fae1b69cafb5f2697e92dfcdee"} Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.605065 4727 scope.go:117] "RemoveContainer" containerID="6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.605285 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.656755 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.678611 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.690887 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:55 crc kubenswrapper[4727]: E0929 10:44:55.691532 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="setup-container" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.691559 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="setup-container" Sep 29 10:44:55 crc kubenswrapper[4727]: E0929 10:44:55.691577 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="rabbitmq" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.691585 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="rabbitmq" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.691808 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="667cfa10-bc94-4788-af5d-296745385383" containerName="rabbitmq" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.693252 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.702214 4727 scope.go:117] "RemoveContainer" containerID="d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.702686 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.702942 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.703138 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.703950 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.704102 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.704275 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8kslj" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.710007 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.719540 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.761074 4727 scope.go:117] "RemoveContainer" containerID="6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7" Sep 29 10:44:55 crc kubenswrapper[4727]: E0929 10:44:55.761774 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7\": container with ID starting with 6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7 not found: ID does not exist" containerID="6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.761812 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7"} err="failed to get container status \"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7\": rpc error: code = NotFound desc = could not find container \"6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7\": container with ID starting with 6f4d96c64f12a872bfaa92ea2b2345dd088af3b42122360c5231da1f91a539c7 not found: ID does not exist" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.761835 4727 scope.go:117] "RemoveContainer" containerID="d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864" Sep 29 10:44:55 crc kubenswrapper[4727]: E0929 10:44:55.762420 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864\": container with ID starting with d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864 not found: ID does not exist" containerID="d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.762460 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864"} err="failed to get container status \"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864\": rpc error: code = NotFound desc = could not find container \"d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864\": container with ID starting with d388a8f4771d3287b0937967191bd7de89807b519d98c5567d31c7be14357864 not found: ID does not exist" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802617 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802680 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802736 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802766 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802794 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802853 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802892 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802916 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.802963 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.803025 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtclv\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-kube-api-access-jtclv\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.803049 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905107 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905534 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905563 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905607 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905648 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905669 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905707 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905773 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtclv\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-kube-api-access-jtclv\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905800 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905851 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.905876 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.907678 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.910034 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.910081 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.910417 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.910644 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.911266 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.913967 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.914976 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.917014 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.921273 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.930921 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtclv\" (UniqueName: \"kubernetes.io/projected/3cb8e279-f15d-46d1-a9fe-38618ca7cc90-kube-api-access-jtclv\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:55 crc kubenswrapper[4727]: I0929 10:44:55.952219 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"3cb8e279-f15d-46d1-a9fe-38618ca7cc90\") " pod="openstack/rabbitmq-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.023468 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.053575 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111023 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111132 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111162 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111195 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111262 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111295 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111366 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111434 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111493 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh8xp\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111549 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.111585 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret\") pod \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\" (UID: \"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8\") " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.122249 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.126853 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.127929 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.130156 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.136623 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info" (OuterVolumeSpecName: "pod-info") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.140679 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp" (OuterVolumeSpecName: "kube-api-access-wh8xp") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "kube-api-access-wh8xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.142570 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.161779 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.162948 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data" (OuterVolumeSpecName: "config-data") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.219942 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf" (OuterVolumeSpecName: "server-conf") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261219 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261285 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261298 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261309 4727 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-pod-info\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261324 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh8xp\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-kube-api-access-wh8xp\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261346 4727 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261356 4727 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261364 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.261377 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.328744 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.363715 4727 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-server-conf\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.364216 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.422239 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" (UID: "a6c551e2-dad5-47a9-b6a1-1489c7fc3be8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.466768 4727 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.616094 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6c551e2-dad5-47a9-b6a1-1489c7fc3be8","Type":"ContainerDied","Data":"f647d4c0c82889d7d4443a6f86c3f021bd40f960896fda7799a1b92d950e3615"} Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.616163 4727 scope.go:117] "RemoveContainer" containerID="e3f4c3abce76e3e7604247f7d881163b52102d07e2ccfa590484a9575c282bf0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.616118 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.643061 4727 scope.go:117] "RemoveContainer" containerID="1f730413d55de9f8401a7955c284e7d886a5fe384f645986bcdcbe43612ec7b6" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.657491 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.666374 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.688032 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:56 crc kubenswrapper[4727]: E0929 10:44:56.688483 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="setup-container" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.688504 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="setup-container" Sep 29 10:44:56 crc kubenswrapper[4727]: E0929 10:44:56.688535 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="rabbitmq" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.688544 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="rabbitmq" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.688731 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" containerName="rabbitmq" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.689840 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694497 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694645 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694681 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694718 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694844 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694913 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-48nph" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.694963 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.704528 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.757948 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877356 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f320dfce-50ce-41ea-b18f-137cd39add94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877755 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877799 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877857 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877886 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877951 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsf8b\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-kube-api-access-gsf8b\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.877983 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.878100 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.878161 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.878295 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.878369 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f320dfce-50ce-41ea-b18f-137cd39add94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979695 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsf8b\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-kube-api-access-gsf8b\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979748 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979777 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979799 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979838 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979868 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f320dfce-50ce-41ea-b18f-137cd39add94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979909 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f320dfce-50ce-41ea-b18f-137cd39add94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979932 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979957 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.979991 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.980008 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.980319 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.980575 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.980643 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.981895 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.982238 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.982901 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f320dfce-50ce-41ea-b18f-137cd39add94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.985947 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f320dfce-50ce-41ea-b18f-137cd39add94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.985966 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.986040 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f320dfce-50ce-41ea-b18f-137cd39add94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:56 crc kubenswrapper[4727]: I0929 10:44:56.986193 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.004124 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsf8b\" (UniqueName: \"kubernetes.io/projected/f320dfce-50ce-41ea-b18f-137cd39add94-kube-api-access-gsf8b\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.015741 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f320dfce-50ce-41ea-b18f-137cd39add94\") " pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.118940 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667cfa10-bc94-4788-af5d-296745385383" path="/var/lib/kubelet/pods/667cfa10-bc94-4788-af5d-296745385383/volumes" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.119711 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6c551e2-dad5-47a9-b6a1-1489c7fc3be8" path="/var/lib/kubelet/pods/a6c551e2-dad5-47a9-b6a1-1489c7fc3be8/volumes" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.317780 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.627089 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cb8e279-f15d-46d1-a9fe-38618ca7cc90","Type":"ContainerStarted","Data":"1d1a32755abfc373593274d1ab09c47638db4e2689b690ea98790d6cf45d8fcb"} Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.627468 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cb8e279-f15d-46d1-a9fe-38618ca7cc90","Type":"ContainerStarted","Data":"7ae9dbf2506dad4408b4771d00dcbdcd6c80484ac6374a748d478461d70912cd"} Sep 29 10:44:57 crc kubenswrapper[4727]: I0929 10:44:57.778688 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:58.638249 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f320dfce-50ce-41ea-b18f-137cd39add94","Type":"ContainerStarted","Data":"d0b096d84f038df26eac974a2219a5377d7059bc45ccc75f0c4801a61e453ee6"} Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:58.638534 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f320dfce-50ce-41ea-b18f-137cd39add94","Type":"ContainerStarted","Data":"1bf46a4669b46c4300cad0937775054122590955e4c73e127c8908b933dfe286"} Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.729313 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.731097 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.736534 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.744412 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.851750 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852024 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852150 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jftxh\" (UniqueName: \"kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852289 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852436 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852686 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.852867 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954097 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954447 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954482 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954507 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954565 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jftxh\" (UniqueName: \"kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954639 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.954670 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955100 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955238 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955430 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955578 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955909 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.955970 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:44:59 crc kubenswrapper[4727]: I0929 10:44:59.974882 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jftxh\" (UniqueName: \"kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh\") pod \"dnsmasq-dns-7877c86b6c-578lv\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.049817 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.132487 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w"] Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.133825 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.144815 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.145023 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.147359 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w"] Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.266602 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.266990 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmb6t\" (UniqueName: \"kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.267036 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.369073 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.369143 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmb6t\" (UniqueName: \"kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.369186 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.370291 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.377629 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.385955 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmb6t\" (UniqueName: \"kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t\") pod \"collect-profiles-29319045-zzl2w\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.496940 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.527222 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:45:00 crc kubenswrapper[4727]: W0929 10:45:00.529888 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7babe14d_2f3a_4a33_9099_c446322fc8ed.slice/crio-2b2aa4915afe2b405f9446e4304d33d29c01ff50b88f45ba13433e7d4b0db3a2 WatchSource:0}: Error finding container 2b2aa4915afe2b405f9446e4304d33d29c01ff50b88f45ba13433e7d4b0db3a2: Status 404 returned error can't find the container with id 2b2aa4915afe2b405f9446e4304d33d29c01ff50b88f45ba13433e7d4b0db3a2 Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.685193 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" event={"ID":"7babe14d-2f3a-4a33-9099-c446322fc8ed","Type":"ContainerStarted","Data":"2b2aa4915afe2b405f9446e4304d33d29c01ff50b88f45ba13433e7d4b0db3a2"} Sep 29 10:45:00 crc kubenswrapper[4727]: I0929 10:45:00.954808 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w"] Sep 29 10:45:00 crc kubenswrapper[4727]: W0929 10:45:00.958091 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61d1abf6_0c7c_422a_90c9_2e5c9b888061.slice/crio-8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469 WatchSource:0}: Error finding container 8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469: Status 404 returned error can't find the container with id 8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469 Sep 29 10:45:01 crc kubenswrapper[4727]: I0929 10:45:01.695859 4727 generic.go:334] "Generic (PLEG): container finished" podID="61d1abf6-0c7c-422a-90c9-2e5c9b888061" containerID="f849a508f8c6124d0eaba17ac34eed0c93600bcd61795235abd655cb3ef700c9" exitCode=0 Sep 29 10:45:01 crc kubenswrapper[4727]: I0929 10:45:01.695926 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" event={"ID":"61d1abf6-0c7c-422a-90c9-2e5c9b888061","Type":"ContainerDied","Data":"f849a508f8c6124d0eaba17ac34eed0c93600bcd61795235abd655cb3ef700c9"} Sep 29 10:45:01 crc kubenswrapper[4727]: I0929 10:45:01.697243 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" event={"ID":"61d1abf6-0c7c-422a-90c9-2e5c9b888061","Type":"ContainerStarted","Data":"8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469"} Sep 29 10:45:01 crc kubenswrapper[4727]: I0929 10:45:01.700454 4727 generic.go:334] "Generic (PLEG): container finished" podID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerID="47acc50e02ccdee1e12e71a618715a82a878945cf3603cda5d7f34a16efe9d8b" exitCode=0 Sep 29 10:45:01 crc kubenswrapper[4727]: I0929 10:45:01.700488 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" event={"ID":"7babe14d-2f3a-4a33-9099-c446322fc8ed","Type":"ContainerDied","Data":"47acc50e02ccdee1e12e71a618715a82a878945cf3603cda5d7f34a16efe9d8b"} Sep 29 10:45:02 crc kubenswrapper[4727]: I0929 10:45:02.710408 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" event={"ID":"7babe14d-2f3a-4a33-9099-c446322fc8ed","Type":"ContainerStarted","Data":"e225a791ca56696a5b56f34ec5b698dd6ceb096e212656bb58ba6385be2591bc"} Sep 29 10:45:02 crc kubenswrapper[4727]: I0929 10:45:02.732821 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" podStartSLOduration=3.7327996150000002 podStartE2EDuration="3.732799615s" podCreationTimestamp="2025-09-29 10:44:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:45:02.727868688 +0000 UTC m=+1372.901182050" watchObservedRunningTime="2025-09-29 10:45:02.732799615 +0000 UTC m=+1372.906112967" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.049858 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:03 crc kubenswrapper[4727]: E0929 10:45:03.078519 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd9580_fa48_48d9_bcf1_4009c21d31a2.slice/crio-1c330581306f3471d5c0df87c19df89d6fcdea895cb83d7ff98e4e492ddd1bc2\": RecentStats: unable to find data in memory cache]" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.116895 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume\") pod \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.117366 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume\") pod \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.117480 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmb6t\" (UniqueName: \"kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t\") pod \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\" (UID: \"61d1abf6-0c7c-422a-90c9-2e5c9b888061\") " Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.119172 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume" (OuterVolumeSpecName: "config-volume") pod "61d1abf6-0c7c-422a-90c9-2e5c9b888061" (UID: "61d1abf6-0c7c-422a-90c9-2e5c9b888061"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.125284 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t" (OuterVolumeSpecName: "kube-api-access-cmb6t") pod "61d1abf6-0c7c-422a-90c9-2e5c9b888061" (UID: "61d1abf6-0c7c-422a-90c9-2e5c9b888061"). InnerVolumeSpecName "kube-api-access-cmb6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.125613 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "61d1abf6-0c7c-422a-90c9-2e5c9b888061" (UID: "61d1abf6-0c7c-422a-90c9-2e5c9b888061"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.220238 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61d1abf6-0c7c-422a-90c9-2e5c9b888061-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.220272 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmb6t\" (UniqueName: \"kubernetes.io/projected/61d1abf6-0c7c-422a-90c9-2e5c9b888061-kube-api-access-cmb6t\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.220285 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61d1abf6-0c7c-422a-90c9-2e5c9b888061-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.719363 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" event={"ID":"61d1abf6-0c7c-422a-90c9-2e5c9b888061","Type":"ContainerDied","Data":"8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469"} Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.719706 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8557103e8b0e7c0a539eeca0da607dc670a5577c204afe91463b30cbea971469" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.719731 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:03 crc kubenswrapper[4727]: I0929 10:45:03.719430 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.051917 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.103757 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.103998 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="dnsmasq-dns" containerID="cri-o://d65da3163f083133f75b9a9838103e80ab44ecb5001e8c74a8e416d28cf569e9" gracePeriod=10 Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.297927 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fb577f7d7-w6xqv"] Sep 29 10:45:10 crc kubenswrapper[4727]: E0929 10:45:10.298515 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d1abf6-0c7c-422a-90c9-2e5c9b888061" containerName="collect-profiles" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.298539 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d1abf6-0c7c-422a-90c9-2e5c9b888061" containerName="collect-profiles" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.298784 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d1abf6-0c7c-422a-90c9-2e5c9b888061" containerName="collect-profiles" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.300210 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.343259 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fb577f7d7-w6xqv"] Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461666 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461755 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l544m\" (UniqueName: \"kubernetes.io/projected/9e9ba97f-39de-4e67-a66c-4c7b575f161c-kube-api-access-l544m\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461792 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461830 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-swift-storage-0\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461855 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461920 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-config\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.461975 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-svc\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.577646 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-config\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.577992 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-svc\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578031 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578078 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l544m\" (UniqueName: \"kubernetes.io/projected/9e9ba97f-39de-4e67-a66c-4c7b575f161c-kube-api-access-l544m\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578100 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578137 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-swift-storage-0\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578160 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.578867 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-config\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.579955 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-svc\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.580024 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-nb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.580581 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-dns-swift-storage-0\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.580669 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-openstack-edpm-ipam\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.583398 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e9ba97f-39de-4e67-a66c-4c7b575f161c-ovsdbserver-sb\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.619525 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l544m\" (UniqueName: \"kubernetes.io/projected/9e9ba97f-39de-4e67-a66c-4c7b575f161c-kube-api-access-l544m\") pod \"dnsmasq-dns-6fb577f7d7-w6xqv\" (UID: \"9e9ba97f-39de-4e67-a66c-4c7b575f161c\") " pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.655504 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.798749 4727 generic.go:334] "Generic (PLEG): container finished" podID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerID="d65da3163f083133f75b9a9838103e80ab44ecb5001e8c74a8e416d28cf569e9" exitCode=0 Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.799724 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" event={"ID":"b0e563a1-166c-4155-a5d9-35c5272c089f","Type":"ContainerDied","Data":"d65da3163f083133f75b9a9838103e80ab44ecb5001e8c74a8e416d28cf569e9"} Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.799853 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" event={"ID":"b0e563a1-166c-4155-a5d9-35c5272c089f","Type":"ContainerDied","Data":"e7c080106de0c9dfb8a930e1e927bcf9d1f4b11a50e26b89be06497498612431"} Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.799880 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7c080106de0c9dfb8a930e1e927bcf9d1f4b11a50e26b89be06497498612431" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.816359 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996329 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996436 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996534 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996574 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996666 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss6zd\" (UniqueName: \"kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:10 crc kubenswrapper[4727]: I0929 10:45:10.996777 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc\") pod \"b0e563a1-166c-4155-a5d9-35c5272c089f\" (UID: \"b0e563a1-166c-4155-a5d9-35c5272c089f\") " Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.004160 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd" (OuterVolumeSpecName: "kube-api-access-ss6zd") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "kube-api-access-ss6zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.049756 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.054991 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.059789 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.061174 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.066216 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config" (OuterVolumeSpecName: "config") pod "b0e563a1-166c-4155-a5d9-35c5272c089f" (UID: "b0e563a1-166c-4155-a5d9-35c5272c089f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099354 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099383 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099393 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099402 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099411 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0e563a1-166c-4155-a5d9-35c5272c089f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.099419 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss6zd\" (UniqueName: \"kubernetes.io/projected/b0e563a1-166c-4155-a5d9-35c5272c089f-kube-api-access-ss6zd\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.180433 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fb577f7d7-w6xqv"] Sep 29 10:45:11 crc kubenswrapper[4727]: W0929 10:45:11.183787 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e9ba97f_39de_4e67_a66c_4c7b575f161c.slice/crio-d8c217a3a37bcd24ebbf433c9932f55a5edefef651ae3202354d2d433ca2c18d WatchSource:0}: Error finding container d8c217a3a37bcd24ebbf433c9932f55a5edefef651ae3202354d2d433ca2c18d: Status 404 returned error can't find the container with id d8c217a3a37bcd24ebbf433c9932f55a5edefef651ae3202354d2d433ca2c18d Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.813094 4727 generic.go:334] "Generic (PLEG): container finished" podID="9e9ba97f-39de-4e67-a66c-4c7b575f161c" containerID="768584ad96a5a1f3de1ce13b40be14792d4ff737593478a57a616d7b346320b2" exitCode=0 Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.813454 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85649f948c-cbw9p" Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.814163 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" event={"ID":"9e9ba97f-39de-4e67-a66c-4c7b575f161c","Type":"ContainerDied","Data":"768584ad96a5a1f3de1ce13b40be14792d4ff737593478a57a616d7b346320b2"} Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.814242 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" event={"ID":"9e9ba97f-39de-4e67-a66c-4c7b575f161c","Type":"ContainerStarted","Data":"d8c217a3a37bcd24ebbf433c9932f55a5edefef651ae3202354d2d433ca2c18d"} Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.862101 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:45:11 crc kubenswrapper[4727]: I0929 10:45:11.872993 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85649f948c-cbw9p"] Sep 29 10:45:12 crc kubenswrapper[4727]: I0929 10:45:12.824301 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" event={"ID":"9e9ba97f-39de-4e67-a66c-4c7b575f161c","Type":"ContainerStarted","Data":"d2b84c6c76c4eefb2be274b2f72cf74a317620202be3ddaa0185b88fe397acf4"} Sep 29 10:45:12 crc kubenswrapper[4727]: I0929 10:45:12.824714 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:12 crc kubenswrapper[4727]: I0929 10:45:12.844895 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" podStartSLOduration=2.844872209 podStartE2EDuration="2.844872209s" podCreationTimestamp="2025-09-29 10:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:45:12.839021778 +0000 UTC m=+1383.012335150" watchObservedRunningTime="2025-09-29 10:45:12.844872209 +0000 UTC m=+1383.018185561" Sep 29 10:45:13 crc kubenswrapper[4727]: I0929 10:45:13.123194 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" path="/var/lib/kubelet/pods/b0e563a1-166c-4155-a5d9-35c5272c089f/volumes" Sep 29 10:45:20 crc kubenswrapper[4727]: I0929 10:45:20.657581 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6fb577f7d7-w6xqv" Sep 29 10:45:20 crc kubenswrapper[4727]: I0929 10:45:20.721382 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:45:20 crc kubenswrapper[4727]: I0929 10:45:20.721671 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="dnsmasq-dns" containerID="cri-o://e225a791ca56696a5b56f34ec5b698dd6ceb096e212656bb58ba6385be2591bc" gracePeriod=10 Sep 29 10:45:20 crc kubenswrapper[4727]: I0929 10:45:20.893876 4727 generic.go:334] "Generic (PLEG): container finished" podID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerID="e225a791ca56696a5b56f34ec5b698dd6ceb096e212656bb58ba6385be2591bc" exitCode=0 Sep 29 10:45:20 crc kubenswrapper[4727]: I0929 10:45:20.893924 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" event={"ID":"7babe14d-2f3a-4a33-9099-c446322fc8ed","Type":"ContainerDied","Data":"e225a791ca56696a5b56f34ec5b698dd6ceb096e212656bb58ba6385be2591bc"} Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.207831 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286095 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286148 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286357 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286386 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286423 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jftxh\" (UniqueName: \"kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286533 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.286571 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0\") pod \"7babe14d-2f3a-4a33-9099-c446322fc8ed\" (UID: \"7babe14d-2f3a-4a33-9099-c446322fc8ed\") " Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.298321 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh" (OuterVolumeSpecName: "kube-api-access-jftxh") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "kube-api-access-jftxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.342180 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config" (OuterVolumeSpecName: "config") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.345747 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.349545 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.350434 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.356214 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.370916 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7babe14d-2f3a-4a33-9099-c446322fc8ed" (UID: "7babe14d-2f3a-4a33-9099-c446322fc8ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389289 4727 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-config\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389347 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389365 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389380 4727 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389394 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jftxh\" (UniqueName: \"kubernetes.io/projected/7babe14d-2f3a-4a33-9099-c446322fc8ed-kube-api-access-jftxh\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389405 4727 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.389415 4727 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7babe14d-2f3a-4a33-9099-c446322fc8ed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.905914 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" event={"ID":"7babe14d-2f3a-4a33-9099-c446322fc8ed","Type":"ContainerDied","Data":"2b2aa4915afe2b405f9446e4304d33d29c01ff50b88f45ba13433e7d4b0db3a2"} Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.905975 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877c86b6c-578lv" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.905976 4727 scope.go:117] "RemoveContainer" containerID="e225a791ca56696a5b56f34ec5b698dd6ceb096e212656bb58ba6385be2591bc" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.928479 4727 scope.go:117] "RemoveContainer" containerID="47acc50e02ccdee1e12e71a618715a82a878945cf3603cda5d7f34a16efe9d8b" Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.941436 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:45:21 crc kubenswrapper[4727]: I0929 10:45:21.951634 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7877c86b6c-578lv"] Sep 29 10:45:23 crc kubenswrapper[4727]: I0929 10:45:23.134846 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" path="/var/lib/kubelet/pods/7babe14d-2f3a-4a33-9099-c446322fc8ed/volumes" Sep 29 10:45:27 crc kubenswrapper[4727]: I0929 10:45:27.954876 4727 generic.go:334] "Generic (PLEG): container finished" podID="3cb8e279-f15d-46d1-a9fe-38618ca7cc90" containerID="1d1a32755abfc373593274d1ab09c47638db4e2689b690ea98790d6cf45d8fcb" exitCode=0 Sep 29 10:45:27 crc kubenswrapper[4727]: I0929 10:45:27.954965 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cb8e279-f15d-46d1-a9fe-38618ca7cc90","Type":"ContainerDied","Data":"1d1a32755abfc373593274d1ab09c47638db4e2689b690ea98790d6cf45d8fcb"} Sep 29 10:45:28 crc kubenswrapper[4727]: I0929 10:45:28.971461 4727 generic.go:334] "Generic (PLEG): container finished" podID="f320dfce-50ce-41ea-b18f-137cd39add94" containerID="d0b096d84f038df26eac974a2219a5377d7059bc45ccc75f0c4801a61e453ee6" exitCode=0 Sep 29 10:45:28 crc kubenswrapper[4727]: I0929 10:45:28.971547 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f320dfce-50ce-41ea-b18f-137cd39add94","Type":"ContainerDied","Data":"d0b096d84f038df26eac974a2219a5377d7059bc45ccc75f0c4801a61e453ee6"} Sep 29 10:45:28 crc kubenswrapper[4727]: I0929 10:45:28.975876 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cb8e279-f15d-46d1-a9fe-38618ca7cc90","Type":"ContainerStarted","Data":"bb1b672c1ad0f9f24d604564f2de3ce314894ef3ecae2d71a01fa03f22555701"} Sep 29 10:45:28 crc kubenswrapper[4727]: I0929 10:45:28.976138 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 29 10:45:29 crc kubenswrapper[4727]: I0929 10:45:29.028916 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=34.028896635 podStartE2EDuration="34.028896635s" podCreationTimestamp="2025-09-29 10:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:45:29.021718289 +0000 UTC m=+1399.195031651" watchObservedRunningTime="2025-09-29 10:45:29.028896635 +0000 UTC m=+1399.202209997" Sep 29 10:45:29 crc kubenswrapper[4727]: I0929 10:45:29.987277 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f320dfce-50ce-41ea-b18f-137cd39add94","Type":"ContainerStarted","Data":"c8115fb40fc54ad4fdd1193fd40b45b502919259055d00fb4b9b29f06f1efc71"} Sep 29 10:45:29 crc kubenswrapper[4727]: I0929 10:45:29.987982 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:45:30 crc kubenswrapper[4727]: I0929 10:45:30.016888 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.016868037 podStartE2EDuration="34.016868037s" podCreationTimestamp="2025-09-29 10:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 10:45:30.009658981 +0000 UTC m=+1400.182972343" watchObservedRunningTime="2025-09-29 10:45:30.016868037 +0000 UTC m=+1400.190181409" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.138034 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j"] Sep 29 10:45:33 crc kubenswrapper[4727]: E0929 10:45:33.139720 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="init" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.139810 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="init" Sep 29 10:45:33 crc kubenswrapper[4727]: E0929 10:45:33.139871 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="init" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.139920 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="init" Sep 29 10:45:33 crc kubenswrapper[4727]: E0929 10:45:33.139986 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.140036 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: E0929 10:45:33.140102 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.140171 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.140483 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e563a1-166c-4155-a5d9-35c5272c089f" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.140577 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="7babe14d-2f3a-4a33-9099-c446322fc8ed" containerName="dnsmasq-dns" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.142704 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.145482 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.145805 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.145884 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.145808 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.147536 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j"] Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.247813 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.248608 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lkbk\" (UniqueName: \"kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.248679 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.249247 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.353051 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lkbk\" (UniqueName: \"kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.353520 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.353607 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.353683 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.361933 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.362173 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.374453 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.376162 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lkbk\" (UniqueName: \"kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:33 crc kubenswrapper[4727]: I0929 10:45:33.473432 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:34 crc kubenswrapper[4727]: I0929 10:45:34.040926 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j"] Sep 29 10:45:34 crc kubenswrapper[4727]: W0929 10:45:34.047545 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod874119b7_b7ce_4e87_a254_892df476efb0.slice/crio-4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee WatchSource:0}: Error finding container 4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee: Status 404 returned error can't find the container with id 4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee Sep 29 10:45:35 crc kubenswrapper[4727]: I0929 10:45:35.027707 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" event={"ID":"874119b7-b7ce-4e87-a254-892df476efb0","Type":"ContainerStarted","Data":"4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee"} Sep 29 10:45:44 crc kubenswrapper[4727]: I0929 10:45:44.115227 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" event={"ID":"874119b7-b7ce-4e87-a254-892df476efb0","Type":"ContainerStarted","Data":"8142ac0dfdccea7de9e113471dd8b5b1e6cfe8fa0d306e1c1cd3b9df220f1472"} Sep 29 10:45:44 crc kubenswrapper[4727]: I0929 10:45:44.137548 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" podStartSLOduration=1.720354105 podStartE2EDuration="11.137528383s" podCreationTimestamp="2025-09-29 10:45:33 +0000 UTC" firstStartedPulling="2025-09-29 10:45:34.049212241 +0000 UTC m=+1404.222525603" lastFinishedPulling="2025-09-29 10:45:43.466386519 +0000 UTC m=+1413.639699881" observedRunningTime="2025-09-29 10:45:44.131772784 +0000 UTC m=+1414.305086146" watchObservedRunningTime="2025-09-29 10:45:44.137528383 +0000 UTC m=+1414.310841745" Sep 29 10:45:46 crc kubenswrapper[4727]: I0929 10:45:46.057597 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 29 10:45:47 crc kubenswrapper[4727]: I0929 10:45:47.320566 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 29 10:45:55 crc kubenswrapper[4727]: I0929 10:45:55.231441 4727 generic.go:334] "Generic (PLEG): container finished" podID="874119b7-b7ce-4e87-a254-892df476efb0" containerID="8142ac0dfdccea7de9e113471dd8b5b1e6cfe8fa0d306e1c1cd3b9df220f1472" exitCode=0 Sep 29 10:45:55 crc kubenswrapper[4727]: I0929 10:45:55.231519 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" event={"ID":"874119b7-b7ce-4e87-a254-892df476efb0","Type":"ContainerDied","Data":"8142ac0dfdccea7de9e113471dd8b5b1e6cfe8fa0d306e1c1cd3b9df220f1472"} Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.649248 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.811552 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory\") pod \"874119b7-b7ce-4e87-a254-892df476efb0\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.811628 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle\") pod \"874119b7-b7ce-4e87-a254-892df476efb0\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.811703 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lkbk\" (UniqueName: \"kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk\") pod \"874119b7-b7ce-4e87-a254-892df476efb0\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.811774 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key\") pod \"874119b7-b7ce-4e87-a254-892df476efb0\" (UID: \"874119b7-b7ce-4e87-a254-892df476efb0\") " Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.822228 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "874119b7-b7ce-4e87-a254-892df476efb0" (UID: "874119b7-b7ce-4e87-a254-892df476efb0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.822437 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk" (OuterVolumeSpecName: "kube-api-access-2lkbk") pod "874119b7-b7ce-4e87-a254-892df476efb0" (UID: "874119b7-b7ce-4e87-a254-892df476efb0"). InnerVolumeSpecName "kube-api-access-2lkbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.838577 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "874119b7-b7ce-4e87-a254-892df476efb0" (UID: "874119b7-b7ce-4e87-a254-892df476efb0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.840324 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory" (OuterVolumeSpecName: "inventory") pod "874119b7-b7ce-4e87-a254-892df476efb0" (UID: "874119b7-b7ce-4e87-a254-892df476efb0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.913666 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.913982 4727 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.914107 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lkbk\" (UniqueName: \"kubernetes.io/projected/874119b7-b7ce-4e87-a254-892df476efb0-kube-api-access-2lkbk\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:56 crc kubenswrapper[4727]: I0929 10:45:56.914184 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/874119b7-b7ce-4e87-a254-892df476efb0-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.251126 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" event={"ID":"874119b7-b7ce-4e87-a254-892df476efb0","Type":"ContainerDied","Data":"4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee"} Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.251474 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ad53d793710a217a319f32fa225f9bd28920790b3dcdf23f2fa50b7e03353ee" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.251203 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.321112 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8"] Sep 29 10:45:57 crc kubenswrapper[4727]: E0929 10:45:57.321618 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874119b7-b7ce-4e87-a254-892df476efb0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.321636 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="874119b7-b7ce-4e87-a254-892df476efb0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.321875 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="874119b7-b7ce-4e87-a254-892df476efb0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.322618 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.328714 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.328739 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.328845 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.335105 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.339602 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8"] Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.424375 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swbgt\" (UniqueName: \"kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.424856 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.425016 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.526636 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.527001 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swbgt\" (UniqueName: \"kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.527091 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.534539 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.541672 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.547163 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swbgt\" (UniqueName: \"kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zbjd8\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:57 crc kubenswrapper[4727]: I0929 10:45:57.663851 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:45:58 crc kubenswrapper[4727]: I0929 10:45:58.223867 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8"] Sep 29 10:45:58 crc kubenswrapper[4727]: I0929 10:45:58.260764 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" event={"ID":"eac8b453-a168-458e-9e0b-28909a6323c4","Type":"ContainerStarted","Data":"216a344dce5f57a75f1c99ca2a681427d63b00695e2963e168afe6b6acc217b8"} Sep 29 10:45:59 crc kubenswrapper[4727]: I0929 10:45:59.276150 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" event={"ID":"eac8b453-a168-458e-9e0b-28909a6323c4","Type":"ContainerStarted","Data":"ec944551fb7a6fa6bf7bfc8fe71b7090a42d0eb349d18d875e48c24e1ad702cb"} Sep 29 10:45:59 crc kubenswrapper[4727]: I0929 10:45:59.306983 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" podStartSLOduration=1.896194429 podStartE2EDuration="2.306910829s" podCreationTimestamp="2025-09-29 10:45:57 +0000 UTC" firstStartedPulling="2025-09-29 10:45:58.233113653 +0000 UTC m=+1428.406427015" lastFinishedPulling="2025-09-29 10:45:58.643830053 +0000 UTC m=+1428.817143415" observedRunningTime="2025-09-29 10:45:59.292692502 +0000 UTC m=+1429.466005874" watchObservedRunningTime="2025-09-29 10:45:59.306910829 +0000 UTC m=+1429.480224191" Sep 29 10:46:02 crc kubenswrapper[4727]: I0929 10:46:02.306394 4727 generic.go:334] "Generic (PLEG): container finished" podID="eac8b453-a168-458e-9e0b-28909a6323c4" containerID="ec944551fb7a6fa6bf7bfc8fe71b7090a42d0eb349d18d875e48c24e1ad702cb" exitCode=0 Sep 29 10:46:02 crc kubenswrapper[4727]: I0929 10:46:02.306515 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" event={"ID":"eac8b453-a168-458e-9e0b-28909a6323c4","Type":"ContainerDied","Data":"ec944551fb7a6fa6bf7bfc8fe71b7090a42d0eb349d18d875e48c24e1ad702cb"} Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.690651 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.857283 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key\") pod \"eac8b453-a168-458e-9e0b-28909a6323c4\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.857527 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory\") pod \"eac8b453-a168-458e-9e0b-28909a6323c4\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.857607 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swbgt\" (UniqueName: \"kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt\") pod \"eac8b453-a168-458e-9e0b-28909a6323c4\" (UID: \"eac8b453-a168-458e-9e0b-28909a6323c4\") " Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.865564 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt" (OuterVolumeSpecName: "kube-api-access-swbgt") pod "eac8b453-a168-458e-9e0b-28909a6323c4" (UID: "eac8b453-a168-458e-9e0b-28909a6323c4"). InnerVolumeSpecName "kube-api-access-swbgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.891312 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eac8b453-a168-458e-9e0b-28909a6323c4" (UID: "eac8b453-a168-458e-9e0b-28909a6323c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.894283 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory" (OuterVolumeSpecName: "inventory") pod "eac8b453-a168-458e-9e0b-28909a6323c4" (UID: "eac8b453-a168-458e-9e0b-28909a6323c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.959779 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.959822 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac8b453-a168-458e-9e0b-28909a6323c4-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:03 crc kubenswrapper[4727]: I0929 10:46:03.959833 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swbgt\" (UniqueName: \"kubernetes.io/projected/eac8b453-a168-458e-9e0b-28909a6323c4-kube-api-access-swbgt\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.325375 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" event={"ID":"eac8b453-a168-458e-9e0b-28909a6323c4","Type":"ContainerDied","Data":"216a344dce5f57a75f1c99ca2a681427d63b00695e2963e168afe6b6acc217b8"} Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.325461 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216a344dce5f57a75f1c99ca2a681427d63b00695e2963e168afe6b6acc217b8" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.325530 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zbjd8" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.411110 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl"] Sep 29 10:46:04 crc kubenswrapper[4727]: E0929 10:46:04.411617 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac8b453-a168-458e-9e0b-28909a6323c4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.411912 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac8b453-a168-458e-9e0b-28909a6323c4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.412113 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac8b453-a168-458e-9e0b-28909a6323c4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.412923 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.415596 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.415690 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.415860 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.415958 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.426942 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl"] Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.569689 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgkp4\" (UniqueName: \"kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.569755 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.569817 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.570091 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.671920 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.672012 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgkp4\" (UniqueName: \"kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.672055 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.672122 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.677992 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.678009 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.685775 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.691240 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgkp4\" (UniqueName: \"kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:04 crc kubenswrapper[4727]: I0929 10:46:04.739370 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:46:05 crc kubenswrapper[4727]: I0929 10:46:05.253832 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl"] Sep 29 10:46:05 crc kubenswrapper[4727]: I0929 10:46:05.334020 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" event={"ID":"21709a4b-d29f-47a8-acf8-ba9c1e2415c3","Type":"ContainerStarted","Data":"ee6453e7dd3b740dbc33c902c8eacb5c4315480bc000185942b09dba17af0537"} Sep 29 10:46:09 crc kubenswrapper[4727]: I0929 10:46:09.374452 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" event={"ID":"21709a4b-d29f-47a8-acf8-ba9c1e2415c3","Type":"ContainerStarted","Data":"d2c517631d794959054273340115113194c32021b7f13055287686693453ac40"} Sep 29 10:46:09 crc kubenswrapper[4727]: I0929 10:46:09.396196 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" podStartSLOduration=1.9133378589999999 podStartE2EDuration="5.396173545s" podCreationTimestamp="2025-09-29 10:46:04 +0000 UTC" firstStartedPulling="2025-09-29 10:46:05.257170211 +0000 UTC m=+1435.430483573" lastFinishedPulling="2025-09-29 10:46:08.740005897 +0000 UTC m=+1438.913319259" observedRunningTime="2025-09-29 10:46:09.389948044 +0000 UTC m=+1439.563261406" watchObservedRunningTime="2025-09-29 10:46:09.396173545 +0000 UTC m=+1439.569486907" Sep 29 10:46:14 crc kubenswrapper[4727]: I0929 10:46:14.107304 4727 scope.go:117] "RemoveContainer" containerID="3bcc578aa5ae896207b924981b8d78582d68cbd8e56462debf537d5dcad10772" Sep 29 10:46:14 crc kubenswrapper[4727]: I0929 10:46:14.140593 4727 scope.go:117] "RemoveContainer" containerID="1fc13bdbedea3ddbc2b9e60abc5c887146237a6a2a5d7e340283a3d33ad137cd" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.089295 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.100523 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.100654 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.170326 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjcxc\" (UniqueName: \"kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.170415 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.170782 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.273299 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.273561 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjcxc\" (UniqueName: \"kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.273640 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.273759 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.273958 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.297586 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjcxc\" (UniqueName: \"kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc\") pod \"redhat-operators-8499h\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.420106 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:30 crc kubenswrapper[4727]: W0929 10:46:30.696499 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf15c2d9_aba7_4758_91ac_28f316ef0cf4.slice/crio-5f98c89f15ec97c9886057a4ec2047dcb9bea7047052aeb2808d3fc03a136b65 WatchSource:0}: Error finding container 5f98c89f15ec97c9886057a4ec2047dcb9bea7047052aeb2808d3fc03a136b65: Status 404 returned error can't find the container with id 5f98c89f15ec97c9886057a4ec2047dcb9bea7047052aeb2808d3fc03a136b65 Sep 29 10:46:30 crc kubenswrapper[4727]: I0929 10:46:30.701055 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:31 crc kubenswrapper[4727]: I0929 10:46:31.620549 4727 generic.go:334] "Generic (PLEG): container finished" podID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerID="7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4" exitCode=0 Sep 29 10:46:31 crc kubenswrapper[4727]: I0929 10:46:31.620602 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerDied","Data":"7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4"} Sep 29 10:46:31 crc kubenswrapper[4727]: I0929 10:46:31.622483 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerStarted","Data":"5f98c89f15ec97c9886057a4ec2047dcb9bea7047052aeb2808d3fc03a136b65"} Sep 29 10:46:33 crc kubenswrapper[4727]: I0929 10:46:33.643557 4727 generic.go:334] "Generic (PLEG): container finished" podID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerID="c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b" exitCode=0 Sep 29 10:46:33 crc kubenswrapper[4727]: I0929 10:46:33.643654 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerDied","Data":"c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b"} Sep 29 10:46:36 crc kubenswrapper[4727]: I0929 10:46:36.692125 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerStarted","Data":"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5"} Sep 29 10:46:36 crc kubenswrapper[4727]: I0929 10:46:36.714361 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8499h" podStartSLOduration=2.945519646 podStartE2EDuration="6.714325456s" podCreationTimestamp="2025-09-29 10:46:30 +0000 UTC" firstStartedPulling="2025-09-29 10:46:31.623674706 +0000 UTC m=+1461.796988068" lastFinishedPulling="2025-09-29 10:46:35.392480506 +0000 UTC m=+1465.565793878" observedRunningTime="2025-09-29 10:46:36.713874994 +0000 UTC m=+1466.887188356" watchObservedRunningTime="2025-09-29 10:46:36.714325456 +0000 UTC m=+1466.887638818" Sep 29 10:46:40 crc kubenswrapper[4727]: I0929 10:46:40.420417 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:40 crc kubenswrapper[4727]: I0929 10:46:40.420929 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:40 crc kubenswrapper[4727]: I0929 10:46:40.468812 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:40 crc kubenswrapper[4727]: I0929 10:46:40.771187 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:40 crc kubenswrapper[4727]: I0929 10:46:40.819924 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:42 crc kubenswrapper[4727]: I0929 10:46:42.743352 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8499h" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="registry-server" containerID="cri-o://461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5" gracePeriod=2 Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.222659 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.338407 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjcxc\" (UniqueName: \"kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc\") pod \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.338827 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content\") pod \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.338885 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities\") pod \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\" (UID: \"bf15c2d9-aba7-4758-91ac-28f316ef0cf4\") " Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.340102 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities" (OuterVolumeSpecName: "utilities") pod "bf15c2d9-aba7-4758-91ac-28f316ef0cf4" (UID: "bf15c2d9-aba7-4758-91ac-28f316ef0cf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.347153 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc" (OuterVolumeSpecName: "kube-api-access-bjcxc") pod "bf15c2d9-aba7-4758-91ac-28f316ef0cf4" (UID: "bf15c2d9-aba7-4758-91ac-28f316ef0cf4"). InnerVolumeSpecName "kube-api-access-bjcxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.433552 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf15c2d9-aba7-4758-91ac-28f316ef0cf4" (UID: "bf15c2d9-aba7-4758-91ac-28f316ef0cf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.441456 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjcxc\" (UniqueName: \"kubernetes.io/projected/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-kube-api-access-bjcxc\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.441508 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.441522 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf15c2d9-aba7-4758-91ac-28f316ef0cf4-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.762984 4727 generic.go:334] "Generic (PLEG): container finished" podID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerID="461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5" exitCode=0 Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.763065 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerDied","Data":"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5"} Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.763143 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8499h" event={"ID":"bf15c2d9-aba7-4758-91ac-28f316ef0cf4","Type":"ContainerDied","Data":"5f98c89f15ec97c9886057a4ec2047dcb9bea7047052aeb2808d3fc03a136b65"} Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.763185 4727 scope.go:117] "RemoveContainer" containerID="461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.763466 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8499h" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.806532 4727 scope.go:117] "RemoveContainer" containerID="c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.811505 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.819907 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8499h"] Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.834598 4727 scope.go:117] "RemoveContainer" containerID="7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.896783 4727 scope.go:117] "RemoveContainer" containerID="461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5" Sep 29 10:46:43 crc kubenswrapper[4727]: E0929 10:46:43.897486 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5\": container with ID starting with 461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5 not found: ID does not exist" containerID="461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.897566 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5"} err="failed to get container status \"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5\": rpc error: code = NotFound desc = could not find container \"461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5\": container with ID starting with 461343884ea4fa11f773a09f41d2f54ac718f6aaf8890344980b09d5b3367bf5 not found: ID does not exist" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.897597 4727 scope.go:117] "RemoveContainer" containerID="c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b" Sep 29 10:46:43 crc kubenswrapper[4727]: E0929 10:46:43.898047 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b\": container with ID starting with c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b not found: ID does not exist" containerID="c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.898074 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b"} err="failed to get container status \"c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b\": rpc error: code = NotFound desc = could not find container \"c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b\": container with ID starting with c0ec58745036f21dac60cd60d37f14ad2c95bcb9c9f01ca55a62638e2e84095b not found: ID does not exist" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.898091 4727 scope.go:117] "RemoveContainer" containerID="7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4" Sep 29 10:46:43 crc kubenswrapper[4727]: E0929 10:46:43.898722 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4\": container with ID starting with 7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4 not found: ID does not exist" containerID="7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4" Sep 29 10:46:43 crc kubenswrapper[4727]: I0929 10:46:43.898744 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4"} err="failed to get container status \"7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4\": rpc error: code = NotFound desc = could not find container \"7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4\": container with ID starting with 7656cf1f7c37b83776013c67f76d06bfa7cc5b331e789c3ff864b5ed276b15d4 not found: ID does not exist" Sep 29 10:46:45 crc kubenswrapper[4727]: I0929 10:46:45.121999 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" path="/var/lib/kubelet/pods/bf15c2d9-aba7-4758-91ac-28f316ef0cf4/volumes" Sep 29 10:46:49 crc kubenswrapper[4727]: I0929 10:46:49.246285 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:46:49 crc kubenswrapper[4727]: I0929 10:46:49.247000 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:47:14 crc kubenswrapper[4727]: I0929 10:47:14.254606 4727 scope.go:117] "RemoveContainer" containerID="a4a67158a9018c3c839f8be630f61699778ccc2313ec80bda6c820a1124b40af" Sep 29 10:47:14 crc kubenswrapper[4727]: I0929 10:47:14.286823 4727 scope.go:117] "RemoveContainer" containerID="6dcc88aeb582b1378f9505c78dea27c8a43e5e2b9d8985064f7dc746630758d1" Sep 29 10:47:14 crc kubenswrapper[4727]: I0929 10:47:14.309324 4727 scope.go:117] "RemoveContainer" containerID="7d2d0fdfae3d791d6f001d70e5b1e8c4547e642a70a19dbd15b5314a2b9d92d5" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.808889 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:17 crc kubenswrapper[4727]: E0929 10:47:17.810008 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="extract-utilities" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.810023 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="extract-utilities" Sep 29 10:47:17 crc kubenswrapper[4727]: E0929 10:47:17.810046 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="registry-server" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.810052 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="registry-server" Sep 29 10:47:17 crc kubenswrapper[4727]: E0929 10:47:17.810083 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="extract-content" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.810091 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="extract-content" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.810264 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf15c2d9-aba7-4758-91ac-28f316ef0cf4" containerName="registry-server" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.811929 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.842811 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.931680 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.931765 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:17 crc kubenswrapper[4727]: I0929 10:47:17.931788 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6dx5\" (UniqueName: \"kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.033561 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.033643 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.033663 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6dx5\" (UniqueName: \"kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.034362 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.034385 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.054440 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6dx5\" (UniqueName: \"kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5\") pod \"redhat-marketplace-85gzj\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.141686 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:18 crc kubenswrapper[4727]: I0929 10:47:18.600043 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:19 crc kubenswrapper[4727]: I0929 10:47:19.125751 4727 generic.go:334] "Generic (PLEG): container finished" podID="ba7ea769-5477-4777-8900-f324b8224bee" containerID="c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407" exitCode=0 Sep 29 10:47:19 crc kubenswrapper[4727]: I0929 10:47:19.125833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerDied","Data":"c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407"} Sep 29 10:47:19 crc kubenswrapper[4727]: I0929 10:47:19.126036 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerStarted","Data":"3a2ceb493dc3148b012a33c0113870954d8a6b3b46f164ba068f0c4f9d634eaa"} Sep 29 10:47:19 crc kubenswrapper[4727]: I0929 10:47:19.286707 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:47:19 crc kubenswrapper[4727]: I0929 10:47:19.286823 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:47:20 crc kubenswrapper[4727]: I0929 10:47:20.140398 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerStarted","Data":"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558"} Sep 29 10:47:21 crc kubenswrapper[4727]: I0929 10:47:21.153976 4727 generic.go:334] "Generic (PLEG): container finished" podID="ba7ea769-5477-4777-8900-f324b8224bee" containerID="01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558" exitCode=0 Sep 29 10:47:21 crc kubenswrapper[4727]: I0929 10:47:21.154101 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerDied","Data":"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558"} Sep 29 10:47:22 crc kubenswrapper[4727]: I0929 10:47:22.165192 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerStarted","Data":"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e"} Sep 29 10:47:22 crc kubenswrapper[4727]: I0929 10:47:22.185732 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-85gzj" podStartSLOduration=2.742901301 podStartE2EDuration="5.185711123s" podCreationTimestamp="2025-09-29 10:47:17 +0000 UTC" firstStartedPulling="2025-09-29 10:47:19.127349199 +0000 UTC m=+1509.300662561" lastFinishedPulling="2025-09-29 10:47:21.570159011 +0000 UTC m=+1511.743472383" observedRunningTime="2025-09-29 10:47:22.183538237 +0000 UTC m=+1512.356851599" watchObservedRunningTime="2025-09-29 10:47:22.185711123 +0000 UTC m=+1512.359024485" Sep 29 10:47:28 crc kubenswrapper[4727]: I0929 10:47:28.142022 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:28 crc kubenswrapper[4727]: I0929 10:47:28.142635 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:28 crc kubenswrapper[4727]: I0929 10:47:28.188883 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:28 crc kubenswrapper[4727]: I0929 10:47:28.273354 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:28 crc kubenswrapper[4727]: I0929 10:47:28.448208 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.247172 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-85gzj" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="registry-server" containerID="cri-o://7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e" gracePeriod=2 Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.765754 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.917399 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities\") pod \"ba7ea769-5477-4777-8900-f324b8224bee\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.917721 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content\") pod \"ba7ea769-5477-4777-8900-f324b8224bee\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.917778 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6dx5\" (UniqueName: \"kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5\") pod \"ba7ea769-5477-4777-8900-f324b8224bee\" (UID: \"ba7ea769-5477-4777-8900-f324b8224bee\") " Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.918448 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities" (OuterVolumeSpecName: "utilities") pod "ba7ea769-5477-4777-8900-f324b8224bee" (UID: "ba7ea769-5477-4777-8900-f324b8224bee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.925756 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5" (OuterVolumeSpecName: "kube-api-access-v6dx5") pod "ba7ea769-5477-4777-8900-f324b8224bee" (UID: "ba7ea769-5477-4777-8900-f324b8224bee"). InnerVolumeSpecName "kube-api-access-v6dx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:47:30 crc kubenswrapper[4727]: I0929 10:47:30.933035 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba7ea769-5477-4777-8900-f324b8224bee" (UID: "ba7ea769-5477-4777-8900-f324b8224bee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.019833 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.019880 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6dx5\" (UniqueName: \"kubernetes.io/projected/ba7ea769-5477-4777-8900-f324b8224bee-kube-api-access-v6dx5\") on node \"crc\" DevicePath \"\"" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.019896 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba7ea769-5477-4777-8900-f324b8224bee-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.258273 4727 generic.go:334] "Generic (PLEG): container finished" podID="ba7ea769-5477-4777-8900-f324b8224bee" containerID="7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e" exitCode=0 Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.258396 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85gzj" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.258472 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerDied","Data":"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e"} Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.258566 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85gzj" event={"ID":"ba7ea769-5477-4777-8900-f324b8224bee","Type":"ContainerDied","Data":"3a2ceb493dc3148b012a33c0113870954d8a6b3b46f164ba068f0c4f9d634eaa"} Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.258592 4727 scope.go:117] "RemoveContainer" containerID="7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.277468 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.286404 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-85gzj"] Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.287405 4727 scope.go:117] "RemoveContainer" containerID="01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.324065 4727 scope.go:117] "RemoveContainer" containerID="c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.359625 4727 scope.go:117] "RemoveContainer" containerID="7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e" Sep 29 10:47:31 crc kubenswrapper[4727]: E0929 10:47:31.360145 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e\": container with ID starting with 7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e not found: ID does not exist" containerID="7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.360184 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e"} err="failed to get container status \"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e\": rpc error: code = NotFound desc = could not find container \"7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e\": container with ID starting with 7edcea5fca9d0cce6124d75c2a7dbdecfdbd0183ef6260a7d6d817cc58e2d23e not found: ID does not exist" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.360210 4727 scope.go:117] "RemoveContainer" containerID="01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558" Sep 29 10:47:31 crc kubenswrapper[4727]: E0929 10:47:31.360643 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558\": container with ID starting with 01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558 not found: ID does not exist" containerID="01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.360671 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558"} err="failed to get container status \"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558\": rpc error: code = NotFound desc = could not find container \"01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558\": container with ID starting with 01a4b90a4c5b938052cfac40ca23dd9f9bb0809239fa4ed468af5941fe77f558 not found: ID does not exist" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.360689 4727 scope.go:117] "RemoveContainer" containerID="c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407" Sep 29 10:47:31 crc kubenswrapper[4727]: E0929 10:47:31.361097 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407\": container with ID starting with c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407 not found: ID does not exist" containerID="c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407" Sep 29 10:47:31 crc kubenswrapper[4727]: I0929 10:47:31.361130 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407"} err="failed to get container status \"c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407\": rpc error: code = NotFound desc = could not find container \"c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407\": container with ID starting with c191ea9117ec5a1ba71a64d2b03c0852e9eadaceadcdee445f6eddd85dc61407 not found: ID does not exist" Sep 29 10:47:33 crc kubenswrapper[4727]: I0929 10:47:33.123227 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba7ea769-5477-4777-8900-f324b8224bee" path="/var/lib/kubelet/pods/ba7ea769-5477-4777-8900-f324b8224bee/volumes" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.246074 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.246687 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.246730 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.247432 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.247487 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" gracePeriod=600 Sep 29 10:47:49 crc kubenswrapper[4727]: E0929 10:47:49.374677 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.444115 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" exitCode=0 Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.444166 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4"} Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.444203 4727 scope.go:117] "RemoveContainer" containerID="3d41880c27e303a1e5c76236335ad1d25ba6ca69d3b35ad052f02dfa96ee0315" Sep 29 10:47:49 crc kubenswrapper[4727]: I0929 10:47:49.445588 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:47:49 crc kubenswrapper[4727]: E0929 10:47:49.446066 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:48:01 crc kubenswrapper[4727]: I0929 10:48:01.108462 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:48:01 crc kubenswrapper[4727]: E0929 10:48:01.109316 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.147054 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:13 crc kubenswrapper[4727]: E0929 10:48:13.148076 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="extract-utilities" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.148093 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="extract-utilities" Sep 29 10:48:13 crc kubenswrapper[4727]: E0929 10:48:13.148135 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="extract-content" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.148142 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="extract-content" Sep 29 10:48:13 crc kubenswrapper[4727]: E0929 10:48:13.148150 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="registry-server" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.148156 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="registry-server" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.148376 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba7ea769-5477-4777-8900-f324b8224bee" containerName="registry-server" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.149879 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.178089 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.280649 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqjh\" (UniqueName: \"kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.280722 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.280902 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.382330 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.382907 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.383034 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqjh\" (UniqueName: \"kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.382911 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.383404 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.410958 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqjh\" (UniqueName: \"kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh\") pod \"certified-operators-6wmxp\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:13 crc kubenswrapper[4727]: I0929 10:48:13.474293 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:14 crc kubenswrapper[4727]: I0929 10:48:14.069178 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:14 crc kubenswrapper[4727]: I0929 10:48:14.729513 4727 generic.go:334] "Generic (PLEG): container finished" podID="691f3907-cb52-4d93-81ec-f238079c3762" containerID="3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0" exitCode=0 Sep 29 10:48:14 crc kubenswrapper[4727]: I0929 10:48:14.729955 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerDied","Data":"3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0"} Sep 29 10:48:14 crc kubenswrapper[4727]: I0929 10:48:14.729999 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerStarted","Data":"50cb3eb2faf9200deaa67e6f0e070d1fabe5a48500f5bbee967fde7fd06d3316"} Sep 29 10:48:15 crc kubenswrapper[4727]: I0929 10:48:15.747147 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerStarted","Data":"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950"} Sep 29 10:48:16 crc kubenswrapper[4727]: I0929 10:48:16.108497 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:48:16 crc kubenswrapper[4727]: E0929 10:48:16.109229 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:48:16 crc kubenswrapper[4727]: I0929 10:48:16.762369 4727 generic.go:334] "Generic (PLEG): container finished" podID="691f3907-cb52-4d93-81ec-f238079c3762" containerID="7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950" exitCode=0 Sep 29 10:48:16 crc kubenswrapper[4727]: I0929 10:48:16.762439 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerDied","Data":"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950"} Sep 29 10:48:17 crc kubenswrapper[4727]: I0929 10:48:17.776849 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerStarted","Data":"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f"} Sep 29 10:48:17 crc kubenswrapper[4727]: I0929 10:48:17.802845 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6wmxp" podStartSLOduration=2.34995169 podStartE2EDuration="4.802826611s" podCreationTimestamp="2025-09-29 10:48:13 +0000 UTC" firstStartedPulling="2025-09-29 10:48:14.736119822 +0000 UTC m=+1564.909433174" lastFinishedPulling="2025-09-29 10:48:17.188994733 +0000 UTC m=+1567.362308095" observedRunningTime="2025-09-29 10:48:17.79812189 +0000 UTC m=+1567.971435272" watchObservedRunningTime="2025-09-29 10:48:17.802826611 +0000 UTC m=+1567.976139973" Sep 29 10:48:23 crc kubenswrapper[4727]: I0929 10:48:23.474549 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:23 crc kubenswrapper[4727]: I0929 10:48:23.475658 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:23 crc kubenswrapper[4727]: I0929 10:48:23.525383 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:23 crc kubenswrapper[4727]: I0929 10:48:23.889950 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:23 crc kubenswrapper[4727]: I0929 10:48:23.943698 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:25 crc kubenswrapper[4727]: I0929 10:48:25.857320 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6wmxp" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="registry-server" containerID="cri-o://30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f" gracePeriod=2 Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.422403 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.493069 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgqjh\" (UniqueName: \"kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh\") pod \"691f3907-cb52-4d93-81ec-f238079c3762\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.493634 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities\") pod \"691f3907-cb52-4d93-81ec-f238079c3762\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.493700 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content\") pod \"691f3907-cb52-4d93-81ec-f238079c3762\" (UID: \"691f3907-cb52-4d93-81ec-f238079c3762\") " Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.494635 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities" (OuterVolumeSpecName: "utilities") pod "691f3907-cb52-4d93-81ec-f238079c3762" (UID: "691f3907-cb52-4d93-81ec-f238079c3762"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.501747 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh" (OuterVolumeSpecName: "kube-api-access-mgqjh") pod "691f3907-cb52-4d93-81ec-f238079c3762" (UID: "691f3907-cb52-4d93-81ec-f238079c3762"). InnerVolumeSpecName "kube-api-access-mgqjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.550055 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "691f3907-cb52-4d93-81ec-f238079c3762" (UID: "691f3907-cb52-4d93-81ec-f238079c3762"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.596052 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgqjh\" (UniqueName: \"kubernetes.io/projected/691f3907-cb52-4d93-81ec-f238079c3762-kube-api-access-mgqjh\") on node \"crc\" DevicePath \"\"" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.596190 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.596202 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691f3907-cb52-4d93-81ec-f238079c3762-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.867776 4727 generic.go:334] "Generic (PLEG): container finished" podID="691f3907-cb52-4d93-81ec-f238079c3762" containerID="30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f" exitCode=0 Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.867833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerDied","Data":"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f"} Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.867867 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wmxp" event={"ID":"691f3907-cb52-4d93-81ec-f238079c3762","Type":"ContainerDied","Data":"50cb3eb2faf9200deaa67e6f0e070d1fabe5a48500f5bbee967fde7fd06d3316"} Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.867893 4727 scope.go:117] "RemoveContainer" containerID="30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.868065 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wmxp" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.889858 4727 scope.go:117] "RemoveContainer" containerID="7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.959087 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.967980 4727 scope.go:117] "RemoveContainer" containerID="3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.968092 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6wmxp"] Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.987568 4727 scope.go:117] "RemoveContainer" containerID="30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f" Sep 29 10:48:26 crc kubenswrapper[4727]: E0929 10:48:26.988031 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f\": container with ID starting with 30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f not found: ID does not exist" containerID="30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.988068 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f"} err="failed to get container status \"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f\": rpc error: code = NotFound desc = could not find container \"30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f\": container with ID starting with 30407674eb34ece939ed1e1a64c16833b15be9d48d2648448bf7cec30ddaba6f not found: ID does not exist" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.988097 4727 scope.go:117] "RemoveContainer" containerID="7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950" Sep 29 10:48:26 crc kubenswrapper[4727]: E0929 10:48:26.988431 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950\": container with ID starting with 7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950 not found: ID does not exist" containerID="7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.988480 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950"} err="failed to get container status \"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950\": rpc error: code = NotFound desc = could not find container \"7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950\": container with ID starting with 7855fd72c78dfe650b27071d2ca1cef7f3058a60f1d79a3d70fc3aab3694d950 not found: ID does not exist" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.988517 4727 scope.go:117] "RemoveContainer" containerID="3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0" Sep 29 10:48:26 crc kubenswrapper[4727]: E0929 10:48:26.988981 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0\": container with ID starting with 3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0 not found: ID does not exist" containerID="3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0" Sep 29 10:48:26 crc kubenswrapper[4727]: I0929 10:48:26.989057 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0"} err="failed to get container status \"3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0\": rpc error: code = NotFound desc = could not find container \"3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0\": container with ID starting with 3ec3a1157f1b827fc02ca4013db833495a84c3b1a7f0c1b8be3a3dc5cd1942c0 not found: ID does not exist" Sep 29 10:48:27 crc kubenswrapper[4727]: I0929 10:48:27.120054 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="691f3907-cb52-4d93-81ec-f238079c3762" path="/var/lib/kubelet/pods/691f3907-cb52-4d93-81ec-f238079c3762/volumes" Sep 29 10:48:28 crc kubenswrapper[4727]: I0929 10:48:28.108618 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:48:28 crc kubenswrapper[4727]: E0929 10:48:28.109744 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:48:40 crc kubenswrapper[4727]: I0929 10:48:40.108216 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:48:40 crc kubenswrapper[4727]: E0929 10:48:40.109208 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:48:54 crc kubenswrapper[4727]: I0929 10:48:54.109150 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:48:54 crc kubenswrapper[4727]: E0929 10:48:54.111759 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:49:05 crc kubenswrapper[4727]: I0929 10:49:05.109009 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:49:05 crc kubenswrapper[4727]: E0929 10:49:05.109866 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:49:11 crc kubenswrapper[4727]: I0929 10:49:11.288675 4727 generic.go:334] "Generic (PLEG): container finished" podID="21709a4b-d29f-47a8-acf8-ba9c1e2415c3" containerID="d2c517631d794959054273340115113194c32021b7f13055287686693453ac40" exitCode=0 Sep 29 10:49:11 crc kubenswrapper[4727]: I0929 10:49:11.288804 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" event={"ID":"21709a4b-d29f-47a8-acf8-ba9c1e2415c3","Type":"ContainerDied","Data":"d2c517631d794959054273340115113194c32021b7f13055287686693453ac40"} Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.783846 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.943839 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory\") pod \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.944055 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle\") pod \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.944188 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgkp4\" (UniqueName: \"kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4\") pod \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.944252 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key\") pod \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\" (UID: \"21709a4b-d29f-47a8-acf8-ba9c1e2415c3\") " Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.951714 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4" (OuterVolumeSpecName: "kube-api-access-qgkp4") pod "21709a4b-d29f-47a8-acf8-ba9c1e2415c3" (UID: "21709a4b-d29f-47a8-acf8-ba9c1e2415c3"). InnerVolumeSpecName "kube-api-access-qgkp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.951739 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "21709a4b-d29f-47a8-acf8-ba9c1e2415c3" (UID: "21709a4b-d29f-47a8-acf8-ba9c1e2415c3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.974257 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21709a4b-d29f-47a8-acf8-ba9c1e2415c3" (UID: "21709a4b-d29f-47a8-acf8-ba9c1e2415c3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:49:12 crc kubenswrapper[4727]: I0929 10:49:12.977878 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory" (OuterVolumeSpecName: "inventory") pod "21709a4b-d29f-47a8-acf8-ba9c1e2415c3" (UID: "21709a4b-d29f-47a8-acf8-ba9c1e2415c3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.047996 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.048068 4727 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.048088 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgkp4\" (UniqueName: \"kubernetes.io/projected/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-kube-api-access-qgkp4\") on node \"crc\" DevicePath \"\"" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.048100 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21709a4b-d29f-47a8-acf8-ba9c1e2415c3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.318161 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" event={"ID":"21709a4b-d29f-47a8-acf8-ba9c1e2415c3","Type":"ContainerDied","Data":"ee6453e7dd3b740dbc33c902c8eacb5c4315480bc000185942b09dba17af0537"} Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.318215 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.318220 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee6453e7dd3b740dbc33c902c8eacb5c4315480bc000185942b09dba17af0537" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387233 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7"] Sep 29 10:49:13 crc kubenswrapper[4727]: E0929 10:49:13.387683 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="extract-content" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387696 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="extract-content" Sep 29 10:49:13 crc kubenswrapper[4727]: E0929 10:49:13.387705 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="extract-utilities" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387712 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="extract-utilities" Sep 29 10:49:13 crc kubenswrapper[4727]: E0929 10:49:13.387739 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="registry-server" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387747 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="registry-server" Sep 29 10:49:13 crc kubenswrapper[4727]: E0929 10:49:13.387761 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21709a4b-d29f-47a8-acf8-ba9c1e2415c3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387767 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="21709a4b-d29f-47a8-acf8-ba9c1e2415c3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387944 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="21709a4b-d29f-47a8-acf8-ba9c1e2415c3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.387967 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="691f3907-cb52-4d93-81ec-f238079c3762" containerName="registry-server" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.388754 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.391317 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.391709 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.392710 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.393369 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.405965 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7"] Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.556324 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.556473 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgw2w\" (UniqueName: \"kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.556524 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.658493 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.658619 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.658680 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgw2w\" (UniqueName: \"kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.663626 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.665038 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.682277 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgw2w\" (UniqueName: \"kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:13 crc kubenswrapper[4727]: I0929 10:49:13.707687 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.297259 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7"] Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.314454 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.335725 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" event={"ID":"96f1c41e-9277-4df7-afd8-711f2efabcf6","Type":"ContainerStarted","Data":"fdf0a0e60fa07ae79d5682d20daf5c44d84cafde1b952ec6ab767456608dc269"} Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.451216 4727 scope.go:117] "RemoveContainer" containerID="67c7252cec796a7a20f060d7d4f80234872610bfca4c87c262152c5141845837" Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.478034 4727 scope.go:117] "RemoveContainer" containerID="e4a48047a2351b17bc778a297af84bb7aeb210f1138bee254c482168cceb4e76" Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.514187 4727 scope.go:117] "RemoveContainer" containerID="5272d13d7f54b139c514b988b5034dbab548028f5bf1270343b36770cb8d2121" Sep 29 10:49:14 crc kubenswrapper[4727]: I0929 10:49:14.537503 4727 scope.go:117] "RemoveContainer" containerID="ebf50f95321bb3eb023e0906cd29514e263856656a723819d07977c5df164b1e" Sep 29 10:49:16 crc kubenswrapper[4727]: I0929 10:49:16.108730 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:49:16 crc kubenswrapper[4727]: E0929 10:49:16.109877 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:49:16 crc kubenswrapper[4727]: I0929 10:49:16.354640 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" event={"ID":"96f1c41e-9277-4df7-afd8-711f2efabcf6","Type":"ContainerStarted","Data":"ac5d6cedc81ff19b30fc8b64d60979b76d89417d24cc2a1c57002ae457cc225e"} Sep 29 10:49:16 crc kubenswrapper[4727]: I0929 10:49:16.376424 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" podStartSLOduration=2.144044429 podStartE2EDuration="3.376406986s" podCreationTimestamp="2025-09-29 10:49:13 +0000 UTC" firstStartedPulling="2025-09-29 10:49:14.314106072 +0000 UTC m=+1624.487419434" lastFinishedPulling="2025-09-29 10:49:15.546468629 +0000 UTC m=+1625.719781991" observedRunningTime="2025-09-29 10:49:16.369635282 +0000 UTC m=+1626.542948654" watchObservedRunningTime="2025-09-29 10:49:16.376406986 +0000 UTC m=+1626.549720348" Sep 29 10:49:30 crc kubenswrapper[4727]: I0929 10:49:30.109273 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:49:30 crc kubenswrapper[4727]: E0929 10:49:30.110152 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:49:41 crc kubenswrapper[4727]: I0929 10:49:41.115029 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:49:41 crc kubenswrapper[4727]: E0929 10:49:41.115884 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:49:48 crc kubenswrapper[4727]: I0929 10:49:48.049237 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6657k"] Sep 29 10:49:48 crc kubenswrapper[4727]: I0929 10:49:48.060861 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6657k"] Sep 29 10:49:49 crc kubenswrapper[4727]: I0929 10:49:49.121007 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86c42eb0-b98f-4f6a-9107-0ece0511ccb2" path="/var/lib/kubelet/pods/86c42eb0-b98f-4f6a-9107-0ece0511ccb2/volumes" Sep 29 10:49:53 crc kubenswrapper[4727]: I0929 10:49:53.110166 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:49:53 crc kubenswrapper[4727]: E0929 10:49:53.111086 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:04 crc kubenswrapper[4727]: I0929 10:50:04.109351 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:50:04 crc kubenswrapper[4727]: E0929 10:50:04.110151 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:06 crc kubenswrapper[4727]: I0929 10:50:06.027422 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nh7m5"] Sep 29 10:50:06 crc kubenswrapper[4727]: I0929 10:50:06.034880 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-9kxvj"] Sep 29 10:50:06 crc kubenswrapper[4727]: I0929 10:50:06.044038 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nh7m5"] Sep 29 10:50:06 crc kubenswrapper[4727]: I0929 10:50:06.055422 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-9kxvj"] Sep 29 10:50:07 crc kubenswrapper[4727]: I0929 10:50:07.119919 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b6fcc5d-2000-448c-9749-0318107fd488" path="/var/lib/kubelet/pods/0b6fcc5d-2000-448c-9749-0318107fd488/volumes" Sep 29 10:50:07 crc kubenswrapper[4727]: I0929 10:50:07.120476 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b7d9089-b633-4493-8d9a-1a98f8febda0" path="/var/lib/kubelet/pods/5b7d9089-b633-4493-8d9a-1a98f8febda0/volumes" Sep 29 10:50:09 crc kubenswrapper[4727]: I0929 10:50:09.036790 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f1a8-account-create-bbh8m"] Sep 29 10:50:09 crc kubenswrapper[4727]: I0929 10:50:09.047025 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f1a8-account-create-bbh8m"] Sep 29 10:50:09 crc kubenswrapper[4727]: I0929 10:50:09.119493 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd310b2-5548-4eb1-8563-aeb4d7b81f74" path="/var/lib/kubelet/pods/9fd310b2-5548-4eb1-8563-aeb4d7b81f74/volumes" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.602212 4727 scope.go:117] "RemoveContainer" containerID="9c209de6c8dc0aa4cf8fe51b198a321f348fffa9ef67f1a5703e153bf05f5d47" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.625437 4727 scope.go:117] "RemoveContainer" containerID="c0080b1618a82670df46877e015798cc5d03f7ae3f3ca6664d8acdf2cf193ca4" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.668797 4727 scope.go:117] "RemoveContainer" containerID="2277155b0c7fb8ae2aa23f083f62ecae16d7e68f4feaaa5386b0bf4be2b7cb5a" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.711926 4727 scope.go:117] "RemoveContainer" containerID="ad0359a69955155ef944dce607d0eca027d3283393b5e65d8c13989d73036e8e" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.755967 4727 scope.go:117] "RemoveContainer" containerID="d65da3163f083133f75b9a9838103e80ab44ecb5001e8c74a8e416d28cf569e9" Sep 29 10:50:14 crc kubenswrapper[4727]: I0929 10:50:14.807794 4727 scope.go:117] "RemoveContainer" containerID="bb0ad55688a8dddba66eb403888fda93b9303a1392918b9850c51986bf1590b9" Sep 29 10:50:16 crc kubenswrapper[4727]: I0929 10:50:16.109150 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:50:16 crc kubenswrapper[4727]: E0929 10:50:16.110245 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:17 crc kubenswrapper[4727]: I0929 10:50:17.032362 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-81c7-account-create-hqw9d"] Sep 29 10:50:17 crc kubenswrapper[4727]: I0929 10:50:17.041268 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-81c7-account-create-hqw9d"] Sep 29 10:50:17 crc kubenswrapper[4727]: I0929 10:50:17.117917 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3425b874-dcfd-47e3-bf52-4510b6ab77bb" path="/var/lib/kubelet/pods/3425b874-dcfd-47e3-bf52-4510b6ab77bb/volumes" Sep 29 10:50:24 crc kubenswrapper[4727]: I0929 10:50:24.026292 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75ca-account-create-8gcdc"] Sep 29 10:50:24 crc kubenswrapper[4727]: I0929 10:50:24.035055 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-75ca-account-create-8gcdc"] Sep 29 10:50:25 crc kubenswrapper[4727]: I0929 10:50:25.118908 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b4d58c3-da91-44ca-af8b-d542b285332a" path="/var/lib/kubelet/pods/3b4d58c3-da91-44ca-af8b-d542b285332a/volumes" Sep 29 10:50:31 crc kubenswrapper[4727]: I0929 10:50:31.120367 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:50:31 crc kubenswrapper[4727]: E0929 10:50:31.121318 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.028848 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-kmtq9"] Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.038765 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bnj9q"] Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.046734 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6bchk"] Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.054995 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-kmtq9"] Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.063515 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6bchk"] Sep 29 10:50:32 crc kubenswrapper[4727]: I0929 10:50:32.071146 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bnj9q"] Sep 29 10:50:33 crc kubenswrapper[4727]: I0929 10:50:33.123589 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d" path="/var/lib/kubelet/pods/36d866d1-84ff-47a8-8f7c-a5b7c5ac1a9d/volumes" Sep 29 10:50:33 crc kubenswrapper[4727]: I0929 10:50:33.124837 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b33a86-c4c4-469f-8168-0c9a84bce521" path="/var/lib/kubelet/pods/68b33a86-c4c4-469f-8168-0c9a84bce521/volumes" Sep 29 10:50:33 crc kubenswrapper[4727]: I0929 10:50:33.125976 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78290c73-5265-43d0-825d-5b9e3164a754" path="/var/lib/kubelet/pods/78290c73-5265-43d0-825d-5b9e3164a754/volumes" Sep 29 10:50:36 crc kubenswrapper[4727]: I0929 10:50:36.031514 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-vjhq6"] Sep 29 10:50:36 crc kubenswrapper[4727]: I0929 10:50:36.041750 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-vjhq6"] Sep 29 10:50:37 crc kubenswrapper[4727]: I0929 10:50:37.127607 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22" path="/var/lib/kubelet/pods/7ed8645d-bd8d-46b7-8ccf-74ea2ed30e22/volumes" Sep 29 10:50:38 crc kubenswrapper[4727]: I0929 10:50:38.044617 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zl5lc"] Sep 29 10:50:38 crc kubenswrapper[4727]: I0929 10:50:38.053632 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zl5lc"] Sep 29 10:50:39 crc kubenswrapper[4727]: I0929 10:50:39.123842 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a838d423-a730-48fe-b345-9e11d8f290a6" path="/var/lib/kubelet/pods/a838d423-a730-48fe-b345-9e11d8f290a6/volumes" Sep 29 10:50:43 crc kubenswrapper[4727]: I0929 10:50:43.109369 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:50:43 crc kubenswrapper[4727]: E0929 10:50:43.110012 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:47 crc kubenswrapper[4727]: I0929 10:50:47.189271 4727 generic.go:334] "Generic (PLEG): container finished" podID="96f1c41e-9277-4df7-afd8-711f2efabcf6" containerID="ac5d6cedc81ff19b30fc8b64d60979b76d89417d24cc2a1c57002ae457cc225e" exitCode=0 Sep 29 10:50:47 crc kubenswrapper[4727]: I0929 10:50:47.189359 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" event={"ID":"96f1c41e-9277-4df7-afd8-711f2efabcf6","Type":"ContainerDied","Data":"ac5d6cedc81ff19b30fc8b64d60979b76d89417d24cc2a1c57002ae457cc225e"} Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.571265 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.714163 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory\") pod \"96f1c41e-9277-4df7-afd8-711f2efabcf6\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.714256 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key\") pod \"96f1c41e-9277-4df7-afd8-711f2efabcf6\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.714416 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgw2w\" (UniqueName: \"kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w\") pod \"96f1c41e-9277-4df7-afd8-711f2efabcf6\" (UID: \"96f1c41e-9277-4df7-afd8-711f2efabcf6\") " Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.721221 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w" (OuterVolumeSpecName: "kube-api-access-wgw2w") pod "96f1c41e-9277-4df7-afd8-711f2efabcf6" (UID: "96f1c41e-9277-4df7-afd8-711f2efabcf6"). InnerVolumeSpecName "kube-api-access-wgw2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.750144 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "96f1c41e-9277-4df7-afd8-711f2efabcf6" (UID: "96f1c41e-9277-4df7-afd8-711f2efabcf6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.755014 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory" (OuterVolumeSpecName: "inventory") pod "96f1c41e-9277-4df7-afd8-711f2efabcf6" (UID: "96f1c41e-9277-4df7-afd8-711f2efabcf6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.817075 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgw2w\" (UniqueName: \"kubernetes.io/projected/96f1c41e-9277-4df7-afd8-711f2efabcf6-kube-api-access-wgw2w\") on node \"crc\" DevicePath \"\"" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.817109 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:50:48 crc kubenswrapper[4727]: I0929 10:50:48.817118 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96f1c41e-9277-4df7-afd8-711f2efabcf6-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.210726 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" event={"ID":"96f1c41e-9277-4df7-afd8-711f2efabcf6","Type":"ContainerDied","Data":"fdf0a0e60fa07ae79d5682d20daf5c44d84cafde1b952ec6ab767456608dc269"} Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.210781 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdf0a0e60fa07ae79d5682d20daf5c44d84cafde1b952ec6ab767456608dc269" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.210850 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.292461 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k"] Sep 29 10:50:49 crc kubenswrapper[4727]: E0929 10:50:49.294899 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f1c41e-9277-4df7-afd8-711f2efabcf6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.294987 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f1c41e-9277-4df7-afd8-711f2efabcf6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.295236 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="96f1c41e-9277-4df7-afd8-711f2efabcf6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.296019 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.298396 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.299576 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.301968 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.308290 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.315070 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k"] Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.437185 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7qct\" (UniqueName: \"kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.437677 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.437786 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.541251 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7qct\" (UniqueName: \"kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.541409 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.541446 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.546534 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.552185 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.563039 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7qct\" (UniqueName: \"kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zr96k\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:49 crc kubenswrapper[4727]: I0929 10:50:49.621153 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:50:50 crc kubenswrapper[4727]: I0929 10:50:50.177479 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k"] Sep 29 10:50:50 crc kubenswrapper[4727]: I0929 10:50:50.224806 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" event={"ID":"e626458d-df81-47e7-aa33-47f32d67759d","Type":"ContainerStarted","Data":"79969128bd9ed1d374f6f8f2d8adf0ae363cb756c89498df0e86bea69899438b"} Sep 29 10:50:51 crc kubenswrapper[4727]: I0929 10:50:51.236398 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" event={"ID":"e626458d-df81-47e7-aa33-47f32d67759d","Type":"ContainerStarted","Data":"39779ecc7e8dbecf212b8b895b0a1301abcd126e691e461662bc948fce1136b8"} Sep 29 10:50:51 crc kubenswrapper[4727]: I0929 10:50:51.260303 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" podStartSLOduration=1.8577269159999998 podStartE2EDuration="2.26028057s" podCreationTimestamp="2025-09-29 10:50:49 +0000 UTC" firstStartedPulling="2025-09-29 10:50:50.189495809 +0000 UTC m=+1720.362809201" lastFinishedPulling="2025-09-29 10:50:50.592049493 +0000 UTC m=+1720.765362855" observedRunningTime="2025-09-29 10:50:51.253577127 +0000 UTC m=+1721.426890499" watchObservedRunningTime="2025-09-29 10:50:51.26028057 +0000 UTC m=+1721.433593922" Sep 29 10:50:55 crc kubenswrapper[4727]: I0929 10:50:55.030697 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2f68-account-create-9qmm6"] Sep 29 10:50:55 crc kubenswrapper[4727]: I0929 10:50:55.040431 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2f68-account-create-9qmm6"] Sep 29 10:50:55 crc kubenswrapper[4727]: I0929 10:50:55.121910 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab205cd-2a27-40df-a1ca-2698379a1148" path="/var/lib/kubelet/pods/9ab205cd-2a27-40df-a1ca-2698379a1148/volumes" Sep 29 10:50:58 crc kubenswrapper[4727]: I0929 10:50:58.110108 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:50:58 crc kubenswrapper[4727]: E0929 10:50:58.111061 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.032901 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7ab5-account-create-bfxmc"] Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.046082 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-bdc4-account-create-7rr6z"] Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.056693 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7ab5-account-create-bfxmc"] Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.066414 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-bdc4-account-create-7rr6z"] Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.121624 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9030b15b-1cf7-4797-ae8e-b6b0a6137b92" path="/var/lib/kubelet/pods/9030b15b-1cf7-4797-ae8e-b6b0a6137b92/volumes" Sep 29 10:50:59 crc kubenswrapper[4727]: I0929 10:50:59.122616 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b1ec441-5723-4541-8614-435ec4c53b07" path="/var/lib/kubelet/pods/9b1ec441-5723-4541-8614-435ec4c53b07/volumes" Sep 29 10:51:04 crc kubenswrapper[4727]: I0929 10:51:04.031130 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nh4r9"] Sep 29 10:51:04 crc kubenswrapper[4727]: I0929 10:51:04.041028 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nh4r9"] Sep 29 10:51:05 crc kubenswrapper[4727]: I0929 10:51:05.040635 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cx7d2"] Sep 29 10:51:05 crc kubenswrapper[4727]: I0929 10:51:05.048280 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cx7d2"] Sep 29 10:51:05 crc kubenswrapper[4727]: I0929 10:51:05.122599 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fc77a68-5c6e-4325-9846-42b1e4f286ea" path="/var/lib/kubelet/pods/9fc77a68-5c6e-4325-9846-42b1e4f286ea/volumes" Sep 29 10:51:05 crc kubenswrapper[4727]: I0929 10:51:05.123690 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f16d06b5-d52a-48bb-8809-f06e0959ea6e" path="/var/lib/kubelet/pods/f16d06b5-d52a-48bb-8809-f06e0959ea6e/volumes" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.112443 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.116133 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.127938 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.166183 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.166252 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxxwm\" (UniqueName: \"kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.166306 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.269090 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.269145 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxxwm\" (UniqueName: \"kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.269220 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.270121 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.270174 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.296369 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxxwm\" (UniqueName: \"kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm\") pod \"community-operators-rgfng\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:08 crc kubenswrapper[4727]: I0929 10:51:08.453527 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:09 crc kubenswrapper[4727]: I0929 10:51:09.078053 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:09 crc kubenswrapper[4727]: I0929 10:51:09.112011 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:51:09 crc kubenswrapper[4727]: E0929 10:51:09.112424 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:51:09 crc kubenswrapper[4727]: I0929 10:51:09.407253 4727 generic.go:334] "Generic (PLEG): container finished" podID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerID="96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa" exitCode=0 Sep 29 10:51:09 crc kubenswrapper[4727]: I0929 10:51:09.407301 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerDied","Data":"96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa"} Sep 29 10:51:09 crc kubenswrapper[4727]: I0929 10:51:09.407326 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerStarted","Data":"ddb744772afe9a0d093eff96d0fa00c41551ffaa131f4a6ad17cb02511adcb1b"} Sep 29 10:51:11 crc kubenswrapper[4727]: I0929 10:51:11.429266 4727 generic.go:334] "Generic (PLEG): container finished" podID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerID="cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93" exitCode=0 Sep 29 10:51:11 crc kubenswrapper[4727]: I0929 10:51:11.429393 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerDied","Data":"cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93"} Sep 29 10:51:12 crc kubenswrapper[4727]: I0929 10:51:12.445665 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerStarted","Data":"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f"} Sep 29 10:51:12 crc kubenswrapper[4727]: I0929 10:51:12.465188 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rgfng" podStartSLOduration=1.9965917659999999 podStartE2EDuration="4.465166091s" podCreationTimestamp="2025-09-29 10:51:08 +0000 UTC" firstStartedPulling="2025-09-29 10:51:09.409887487 +0000 UTC m=+1739.583200849" lastFinishedPulling="2025-09-29 10:51:11.878461822 +0000 UTC m=+1742.051775174" observedRunningTime="2025-09-29 10:51:12.464620887 +0000 UTC m=+1742.637934269" watchObservedRunningTime="2025-09-29 10:51:12.465166091 +0000 UTC m=+1742.638479453" Sep 29 10:51:14 crc kubenswrapper[4727]: I0929 10:51:14.951379 4727 scope.go:117] "RemoveContainer" containerID="71e7bcd6121589881bbcccf09df7a13ba17446dc9e858f0ebfcacd8579bcc0c9" Sep 29 10:51:14 crc kubenswrapper[4727]: I0929 10:51:14.978690 4727 scope.go:117] "RemoveContainer" containerID="2549db53c32fed963c223d4ee19940e442f6d961fbbc8c9f381f63fd4fa65bd7" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.028675 4727 scope.go:117] "RemoveContainer" containerID="41d2eb378ad10a0657fc3ccafb5692d2abafc83e5b33dda6ea43161e8a49c698" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.075443 4727 scope.go:117] "RemoveContainer" containerID="f08a74b68b3512eab0caf35f07e1f1e08a7a11fe07dc70518c49d4eb79cd8f20" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.127615 4727 scope.go:117] "RemoveContainer" containerID="e65795c49dc721454002fc44688fe81f9664c5b5dcb7d72e0384d0feca42c8e8" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.160048 4727 scope.go:117] "RemoveContainer" containerID="05dd1927ee7df3a348122fc86cee1eac5074b30ee915edf9ddc7d0943d823b89" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.196105 4727 scope.go:117] "RemoveContainer" containerID="def633393858396e91b3eaeebae4c6955672243cac3c887c2a584bb8bb4fe39f" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.213982 4727 scope.go:117] "RemoveContainer" containerID="08b8445010b3cd7b845a00a34267d69e84e4830688094a879c4a39847be1cde4" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.237594 4727 scope.go:117] "RemoveContainer" containerID="6512a5663c0429b948b72b58ed9346f051b8ec842852e16c408b2f4d9b965451" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.265302 4727 scope.go:117] "RemoveContainer" containerID="2cf4f9b396b3e7d89238c61a8c91e2589898f7c6cc75e004dd9467a094110a6f" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.296444 4727 scope.go:117] "RemoveContainer" containerID="e68db5791dc8003f4e619fea8afd301620b3a99fbace6a84e941c3f5b1d36cb0" Sep 29 10:51:15 crc kubenswrapper[4727]: I0929 10:51:15.333298 4727 scope.go:117] "RemoveContainer" containerID="c803ad8f87069d1b6a8b4b56f280c32fd59952bdab95195e3cfd4150cdf6ce00" Sep 29 10:51:18 crc kubenswrapper[4727]: I0929 10:51:18.454499 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:18 crc kubenswrapper[4727]: I0929 10:51:18.456214 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:18 crc kubenswrapper[4727]: I0929 10:51:18.530770 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:19 crc kubenswrapper[4727]: I0929 10:51:19.062515 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-68lgd"] Sep 29 10:51:19 crc kubenswrapper[4727]: I0929 10:51:19.072821 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-68lgd"] Sep 29 10:51:19 crc kubenswrapper[4727]: I0929 10:51:19.123376 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fb3935-9e35-4a18-a2e4-b4442581bc00" path="/var/lib/kubelet/pods/74fb3935-9e35-4a18-a2e4-b4442581bc00/volumes" Sep 29 10:51:19 crc kubenswrapper[4727]: I0929 10:51:19.566241 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:19 crc kubenswrapper[4727]: I0929 10:51:19.623011 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:21 crc kubenswrapper[4727]: I0929 10:51:21.544949 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rgfng" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="registry-server" containerID="cri-o://606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f" gracePeriod=2 Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.046688 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.104737 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities\") pod \"7e9f1eac-d1d1-4e62-af29-0067468733a4\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.104858 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content\") pod \"7e9f1eac-d1d1-4e62-af29-0067468733a4\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.104977 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxxwm\" (UniqueName: \"kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm\") pod \"7e9f1eac-d1d1-4e62-af29-0067468733a4\" (UID: \"7e9f1eac-d1d1-4e62-af29-0067468733a4\") " Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.106280 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities" (OuterVolumeSpecName: "utilities") pod "7e9f1eac-d1d1-4e62-af29-0067468733a4" (UID: "7e9f1eac-d1d1-4e62-af29-0067468733a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.115801 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm" (OuterVolumeSpecName: "kube-api-access-lxxwm") pod "7e9f1eac-d1d1-4e62-af29-0067468733a4" (UID: "7e9f1eac-d1d1-4e62-af29-0067468733a4"). InnerVolumeSpecName "kube-api-access-lxxwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.207064 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxxwm\" (UniqueName: \"kubernetes.io/projected/7e9f1eac-d1d1-4e62-af29-0067468733a4-kube-api-access-lxxwm\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.207116 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.500525 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e9f1eac-d1d1-4e62-af29-0067468733a4" (UID: "7e9f1eac-d1d1-4e62-af29-0067468733a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.514845 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9f1eac-d1d1-4e62-af29-0067468733a4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.558016 4727 generic.go:334] "Generic (PLEG): container finished" podID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerID="606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f" exitCode=0 Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.558112 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgfng" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.558118 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerDied","Data":"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f"} Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.559169 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgfng" event={"ID":"7e9f1eac-d1d1-4e62-af29-0067468733a4","Type":"ContainerDied","Data":"ddb744772afe9a0d093eff96d0fa00c41551ffaa131f4a6ad17cb02511adcb1b"} Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.559198 4727 scope.go:117] "RemoveContainer" containerID="606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.601392 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.603149 4727 scope.go:117] "RemoveContainer" containerID="cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.617602 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rgfng"] Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.635079 4727 scope.go:117] "RemoveContainer" containerID="96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa" Sep 29 10:51:22 crc kubenswrapper[4727]: E0929 10:51:22.661182 4727 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e9f1eac_d1d1_4e62_af29_0067468733a4.slice/crio-ddb744772afe9a0d093eff96d0fa00c41551ffaa131f4a6ad17cb02511adcb1b\": RecentStats: unable to find data in memory cache]" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.673244 4727 scope.go:117] "RemoveContainer" containerID="606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f" Sep 29 10:51:22 crc kubenswrapper[4727]: E0929 10:51:22.674085 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f\": container with ID starting with 606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f not found: ID does not exist" containerID="606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.674215 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f"} err="failed to get container status \"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f\": rpc error: code = NotFound desc = could not find container \"606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f\": container with ID starting with 606cecb52f10a28ba3d5454819b3a9896bd695240702ce82235c34b3d8e59a2f not found: ID does not exist" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.674377 4727 scope.go:117] "RemoveContainer" containerID="cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93" Sep 29 10:51:22 crc kubenswrapper[4727]: E0929 10:51:22.674995 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93\": container with ID starting with cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93 not found: ID does not exist" containerID="cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.675100 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93"} err="failed to get container status \"cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93\": rpc error: code = NotFound desc = could not find container \"cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93\": container with ID starting with cc7ad8b3f0d0fe884717436d2829be393bcb2b05e132e5b70d9c6f2ea22bae93 not found: ID does not exist" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.675173 4727 scope.go:117] "RemoveContainer" containerID="96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa" Sep 29 10:51:22 crc kubenswrapper[4727]: E0929 10:51:22.676103 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa\": container with ID starting with 96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa not found: ID does not exist" containerID="96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa" Sep 29 10:51:22 crc kubenswrapper[4727]: I0929 10:51:22.676257 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa"} err="failed to get container status \"96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa\": rpc error: code = NotFound desc = could not find container \"96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa\": container with ID starting with 96659223b269486ad0e07b6789b245ab0b04b3dd4611a3b882da753685b3c9aa not found: ID does not exist" Sep 29 10:51:23 crc kubenswrapper[4727]: I0929 10:51:23.109113 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:51:23 crc kubenswrapper[4727]: E0929 10:51:23.109468 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:51:23 crc kubenswrapper[4727]: I0929 10:51:23.123417 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" path="/var/lib/kubelet/pods/7e9f1eac-d1d1-4e62-af29-0067468733a4/volumes" Sep 29 10:51:34 crc kubenswrapper[4727]: I0929 10:51:34.108468 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:51:34 crc kubenswrapper[4727]: E0929 10:51:34.109332 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:51:41 crc kubenswrapper[4727]: I0929 10:51:41.047393 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-8n57q"] Sep 29 10:51:41 crc kubenswrapper[4727]: I0929 10:51:41.057384 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-8n57q"] Sep 29 10:51:41 crc kubenswrapper[4727]: I0929 10:51:41.121280 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ab43089-e386-4585-9324-79b4750eeeff" path="/var/lib/kubelet/pods/2ab43089-e386-4585-9324-79b4750eeeff/volumes" Sep 29 10:51:46 crc kubenswrapper[4727]: I0929 10:51:46.108455 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:51:46 crc kubenswrapper[4727]: E0929 10:51:46.109708 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:51:56 crc kubenswrapper[4727]: I0929 10:51:56.040973 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-sg2mp"] Sep 29 10:51:56 crc kubenswrapper[4727]: I0929 10:51:56.058872 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-sg2mp"] Sep 29 10:51:56 crc kubenswrapper[4727]: I0929 10:51:56.943404 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" event={"ID":"e626458d-df81-47e7-aa33-47f32d67759d","Type":"ContainerDied","Data":"39779ecc7e8dbecf212b8b895b0a1301abcd126e691e461662bc948fce1136b8"} Sep 29 10:51:56 crc kubenswrapper[4727]: I0929 10:51:56.943389 4727 generic.go:334] "Generic (PLEG): container finished" podID="e626458d-df81-47e7-aa33-47f32d67759d" containerID="39779ecc7e8dbecf212b8b895b0a1301abcd126e691e461662bc948fce1136b8" exitCode=0 Sep 29 10:51:57 crc kubenswrapper[4727]: I0929 10:51:57.129166 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62cb5972-95c1-4963-b0f2-bfee701d6453" path="/var/lib/kubelet/pods/62cb5972-95c1-4963-b0f2-bfee701d6453/volumes" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.108264 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:51:58 crc kubenswrapper[4727]: E0929 10:51:58.109023 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.490929 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.688469 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory\") pod \"e626458d-df81-47e7-aa33-47f32d67759d\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.688590 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key\") pod \"e626458d-df81-47e7-aa33-47f32d67759d\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.688859 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7qct\" (UniqueName: \"kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct\") pod \"e626458d-df81-47e7-aa33-47f32d67759d\" (UID: \"e626458d-df81-47e7-aa33-47f32d67759d\") " Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.697384 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct" (OuterVolumeSpecName: "kube-api-access-d7qct") pod "e626458d-df81-47e7-aa33-47f32d67759d" (UID: "e626458d-df81-47e7-aa33-47f32d67759d"). InnerVolumeSpecName "kube-api-access-d7qct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.722602 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory" (OuterVolumeSpecName: "inventory") pod "e626458d-df81-47e7-aa33-47f32d67759d" (UID: "e626458d-df81-47e7-aa33-47f32d67759d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.729744 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e626458d-df81-47e7-aa33-47f32d67759d" (UID: "e626458d-df81-47e7-aa33-47f32d67759d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.791697 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7qct\" (UniqueName: \"kubernetes.io/projected/e626458d-df81-47e7-aa33-47f32d67759d-kube-api-access-d7qct\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.791747 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:58 crc kubenswrapper[4727]: I0929 10:51:58.791760 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e626458d-df81-47e7-aa33-47f32d67759d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:58.989702 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" event={"ID":"e626458d-df81-47e7-aa33-47f32d67759d","Type":"ContainerDied","Data":"79969128bd9ed1d374f6f8f2d8adf0ae363cb756c89498df0e86bea69899438b"} Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:58.989755 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79969128bd9ed1d374f6f8f2d8adf0ae363cb756c89498df0e86bea69899438b" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:58.989877 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zr96k" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075066 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c"] Sep 29 10:51:59 crc kubenswrapper[4727]: E0929 10:51:59.075586 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e626458d-df81-47e7-aa33-47f32d67759d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075599 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="e626458d-df81-47e7-aa33-47f32d67759d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:51:59 crc kubenswrapper[4727]: E0929 10:51:59.075616 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="extract-content" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075632 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="extract-content" Sep 29 10:51:59 crc kubenswrapper[4727]: E0929 10:51:59.075649 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="extract-utilities" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075655 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="extract-utilities" Sep 29 10:51:59 crc kubenswrapper[4727]: E0929 10:51:59.075697 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="registry-server" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075707 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="registry-server" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075918 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9f1eac-d1d1-4e62-af29-0067468733a4" containerName="registry-server" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.075937 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="e626458d-df81-47e7-aa33-47f32d67759d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.076752 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.079413 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.079471 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.083576 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.083727 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.086383 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c"] Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.121397 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlkpg\" (UniqueName: \"kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.121634 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.121758 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.223328 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.223466 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.223607 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlkpg\" (UniqueName: \"kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.231260 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.231931 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.249598 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlkpg\" (UniqueName: \"kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.441367 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:51:59 crc kubenswrapper[4727]: I0929 10:51:59.819643 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c"] Sep 29 10:52:00 crc kubenswrapper[4727]: I0929 10:52:00.001989 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" event={"ID":"a2a42153-5fde-492b-9564-fded49a50ebd","Type":"ContainerStarted","Data":"17b28efdea2a9d13f6b930170f8aee6e9e6d6256ddd820eb106927d5777119c4"} Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.037813 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dbvq2"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.047698 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dbvq2"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.058272 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-l6trr"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.069565 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-rrbt8"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.080189 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-rrbt8"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.088310 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-l6trr"] Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.121467 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6be32091-cf3a-4931-bcb8-2c867604774e" path="/var/lib/kubelet/pods/6be32091-cf3a-4931-bcb8-2c867604774e/volumes" Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.122462 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77b2d573-abbb-491f-8939-6c7f79082906" path="/var/lib/kubelet/pods/77b2d573-abbb-491f-8939-6c7f79082906/volumes" Sep 29 10:52:01 crc kubenswrapper[4727]: I0929 10:52:01.124942 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="973c7613-3c61-4534-b26e-e1d0ea649bc9" path="/var/lib/kubelet/pods/973c7613-3c61-4534-b26e-e1d0ea649bc9/volumes" Sep 29 10:52:02 crc kubenswrapper[4727]: I0929 10:52:02.026263 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" event={"ID":"a2a42153-5fde-492b-9564-fded49a50ebd","Type":"ContainerStarted","Data":"636f1be8a93df76440e74cf2215075c7d4cb429b04e3ec2e4631bcc84cfc51a3"} Sep 29 10:52:02 crc kubenswrapper[4727]: I0929 10:52:02.046840 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" podStartSLOduration=1.637074873 podStartE2EDuration="3.046822453s" podCreationTimestamp="2025-09-29 10:51:59 +0000 UTC" firstStartedPulling="2025-09-29 10:51:59.828211908 +0000 UTC m=+1790.001525270" lastFinishedPulling="2025-09-29 10:52:01.237959488 +0000 UTC m=+1791.411272850" observedRunningTime="2025-09-29 10:52:02.046815623 +0000 UTC m=+1792.220129035" watchObservedRunningTime="2025-09-29 10:52:02.046822453 +0000 UTC m=+1792.220135816" Sep 29 10:52:07 crc kubenswrapper[4727]: I0929 10:52:07.051208 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-feac-account-create-j9dr8"] Sep 29 10:52:07 crc kubenswrapper[4727]: I0929 10:52:07.060110 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-feac-account-create-j9dr8"] Sep 29 10:52:07 crc kubenswrapper[4727]: I0929 10:52:07.076061 4727 generic.go:334] "Generic (PLEG): container finished" podID="a2a42153-5fde-492b-9564-fded49a50ebd" containerID="636f1be8a93df76440e74cf2215075c7d4cb429b04e3ec2e4631bcc84cfc51a3" exitCode=0 Sep 29 10:52:07 crc kubenswrapper[4727]: I0929 10:52:07.076144 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" event={"ID":"a2a42153-5fde-492b-9564-fded49a50ebd","Type":"ContainerDied","Data":"636f1be8a93df76440e74cf2215075c7d4cb429b04e3ec2e4631bcc84cfc51a3"} Sep 29 10:52:07 crc kubenswrapper[4727]: I0929 10:52:07.123915 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f04c134-ab93-4a72-9e26-eee995532432" path="/var/lib/kubelet/pods/6f04c134-ab93-4a72-9e26-eee995532432/volumes" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.522306 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.656846 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlkpg\" (UniqueName: \"kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg\") pod \"a2a42153-5fde-492b-9564-fded49a50ebd\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.657049 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key\") pod \"a2a42153-5fde-492b-9564-fded49a50ebd\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.657092 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory\") pod \"a2a42153-5fde-492b-9564-fded49a50ebd\" (UID: \"a2a42153-5fde-492b-9564-fded49a50ebd\") " Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.663945 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg" (OuterVolumeSpecName: "kube-api-access-rlkpg") pod "a2a42153-5fde-492b-9564-fded49a50ebd" (UID: "a2a42153-5fde-492b-9564-fded49a50ebd"). InnerVolumeSpecName "kube-api-access-rlkpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.688790 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a2a42153-5fde-492b-9564-fded49a50ebd" (UID: "a2a42153-5fde-492b-9564-fded49a50ebd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.700666 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory" (OuterVolumeSpecName: "inventory") pod "a2a42153-5fde-492b-9564-fded49a50ebd" (UID: "a2a42153-5fde-492b-9564-fded49a50ebd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.758902 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlkpg\" (UniqueName: \"kubernetes.io/projected/a2a42153-5fde-492b-9564-fded49a50ebd-kube-api-access-rlkpg\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.758933 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:08 crc kubenswrapper[4727]: I0929 10:52:08.758946 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2a42153-5fde-492b-9564-fded49a50ebd-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.100270 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.100133 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c" event={"ID":"a2a42153-5fde-492b-9564-fded49a50ebd","Type":"ContainerDied","Data":"17b28efdea2a9d13f6b930170f8aee6e9e6d6256ddd820eb106927d5777119c4"} Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.101605 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17b28efdea2a9d13f6b930170f8aee6e9e6d6256ddd820eb106927d5777119c4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.259693 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4"] Sep 29 10:52:09 crc kubenswrapper[4727]: E0929 10:52:09.260133 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a42153-5fde-492b-9564-fded49a50ebd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.260154 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a42153-5fde-492b-9564-fded49a50ebd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.260465 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2a42153-5fde-492b-9564-fded49a50ebd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.261219 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.265421 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.265632 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.265757 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.265907 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.272720 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4"] Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.375853 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.376501 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxb6\" (UniqueName: \"kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.376694 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.479008 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.479240 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxb6\" (UniqueName: \"kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.479361 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.490933 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.494173 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.506580 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxb6\" (UniqueName: \"kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-tzhl4\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:09 crc kubenswrapper[4727]: I0929 10:52:09.590962 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:10 crc kubenswrapper[4727]: I0929 10:52:10.155179 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4"] Sep 29 10:52:11 crc kubenswrapper[4727]: I0929 10:52:11.118394 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:52:11 crc kubenswrapper[4727]: E0929 10:52:11.119012 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:52:11 crc kubenswrapper[4727]: I0929 10:52:11.127024 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" event={"ID":"99127a9e-ae57-42b7-9392-e3f3d7ab43ff","Type":"ContainerStarted","Data":"2a6f13066553053520e9cfaa887fab74ec204e3b5220f9d5acf864b58338c878"} Sep 29 10:52:11 crc kubenswrapper[4727]: I0929 10:52:11.717196 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:52:12 crc kubenswrapper[4727]: I0929 10:52:12.137973 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" event={"ID":"99127a9e-ae57-42b7-9392-e3f3d7ab43ff","Type":"ContainerStarted","Data":"b456f112103c9e83e882a142c344831b1b2d1435f1a3c79dcc426aa0a594f5d8"} Sep 29 10:52:12 crc kubenswrapper[4727]: I0929 10:52:12.159233 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" podStartSLOduration=1.614114192 podStartE2EDuration="3.159204254s" podCreationTimestamp="2025-09-29 10:52:09 +0000 UTC" firstStartedPulling="2025-09-29 10:52:10.167953997 +0000 UTC m=+1800.341267369" lastFinishedPulling="2025-09-29 10:52:11.713044069 +0000 UTC m=+1801.886357431" observedRunningTime="2025-09-29 10:52:12.158920077 +0000 UTC m=+1802.332233429" watchObservedRunningTime="2025-09-29 10:52:12.159204254 +0000 UTC m=+1802.332517626" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.544303 4727 scope.go:117] "RemoveContainer" containerID="a05bb2c1ba81f073e36943b5bbad927ad3c500e9367e937bf7845db8d01ad98a" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.587435 4727 scope.go:117] "RemoveContainer" containerID="485e5cf3b6626ea3f28dc564f4341b93555e74bd4bb70dfc6245139ad307c02d" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.637456 4727 scope.go:117] "RemoveContainer" containerID="ed760ac0fa0b366f35c1ecce6501e8be5a490d5ac39a5aac4e37816fdd40d529" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.682923 4727 scope.go:117] "RemoveContainer" containerID="94e3bf00423422c8c394a3220719710323321caddfb62bd0fb53061271bd7f5a" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.710364 4727 scope.go:117] "RemoveContainer" containerID="0a94d38d0c451c27ed76e6ae16ddd0fd73b4aa3755c57a5fc2e041bfa37227fa" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.834011 4727 scope.go:117] "RemoveContainer" containerID="24f9c8d5a43eabb8016a1b0fa2a9548dfe15ff4f232d0bb677913c0c30063dac" Sep 29 10:52:15 crc kubenswrapper[4727]: I0929 10:52:15.867387 4727 scope.go:117] "RemoveContainer" containerID="ff3a5688d5e995bc23e2c566cee352cfc332a17885d01f72318a20848b544e2f" Sep 29 10:52:18 crc kubenswrapper[4727]: I0929 10:52:18.037187 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ad4a-account-create-blgkd"] Sep 29 10:52:18 crc kubenswrapper[4727]: I0929 10:52:18.052726 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f241-account-create-rk5l6"] Sep 29 10:52:18 crc kubenswrapper[4727]: I0929 10:52:18.065044 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f241-account-create-rk5l6"] Sep 29 10:52:18 crc kubenswrapper[4727]: I0929 10:52:18.076583 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ad4a-account-create-blgkd"] Sep 29 10:52:19 crc kubenswrapper[4727]: I0929 10:52:19.126124 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eedcefb-dda0-4045-abbd-611299c470aa" path="/var/lib/kubelet/pods/2eedcefb-dda0-4045-abbd-611299c470aa/volumes" Sep 29 10:52:19 crc kubenswrapper[4727]: I0929 10:52:19.126723 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63978328-20b2-46a4-a423-b9c15c0c03cd" path="/var/lib/kubelet/pods/63978328-20b2-46a4-a423-b9c15c0c03cd/volumes" Sep 29 10:52:25 crc kubenswrapper[4727]: I0929 10:52:25.109399 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:52:25 crc kubenswrapper[4727]: E0929 10:52:25.110108 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:52:37 crc kubenswrapper[4727]: I0929 10:52:37.108452 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:52:37 crc kubenswrapper[4727]: E0929 10:52:37.109457 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:52:46 crc kubenswrapper[4727]: I0929 10:52:46.447230 4727 generic.go:334] "Generic (PLEG): container finished" podID="99127a9e-ae57-42b7-9392-e3f3d7ab43ff" containerID="b456f112103c9e83e882a142c344831b1b2d1435f1a3c79dcc426aa0a594f5d8" exitCode=0 Sep 29 10:52:46 crc kubenswrapper[4727]: I0929 10:52:46.447297 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" event={"ID":"99127a9e-ae57-42b7-9392-e3f3d7ab43ff","Type":"ContainerDied","Data":"b456f112103c9e83e882a142c344831b1b2d1435f1a3c79dcc426aa0a594f5d8"} Sep 29 10:52:47 crc kubenswrapper[4727]: I0929 10:52:47.885992 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.002080 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjxb6\" (UniqueName: \"kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6\") pod \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.002196 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory\") pod \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.002409 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key\") pod \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\" (UID: \"99127a9e-ae57-42b7-9392-e3f3d7ab43ff\") " Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.010985 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6" (OuterVolumeSpecName: "kube-api-access-hjxb6") pod "99127a9e-ae57-42b7-9392-e3f3d7ab43ff" (UID: "99127a9e-ae57-42b7-9392-e3f3d7ab43ff"). InnerVolumeSpecName "kube-api-access-hjxb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.034972 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99127a9e-ae57-42b7-9392-e3f3d7ab43ff" (UID: "99127a9e-ae57-42b7-9392-e3f3d7ab43ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.046231 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory" (OuterVolumeSpecName: "inventory") pod "99127a9e-ae57-42b7-9392-e3f3d7ab43ff" (UID: "99127a9e-ae57-42b7-9392-e3f3d7ab43ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.105707 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.105750 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.105762 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjxb6\" (UniqueName: \"kubernetes.io/projected/99127a9e-ae57-42b7-9392-e3f3d7ab43ff-kube-api-access-hjxb6\") on node \"crc\" DevicePath \"\"" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.475686 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" event={"ID":"99127a9e-ae57-42b7-9392-e3f3d7ab43ff","Type":"ContainerDied","Data":"2a6f13066553053520e9cfaa887fab74ec204e3b5220f9d5acf864b58338c878"} Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.475733 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a6f13066553053520e9cfaa887fab74ec204e3b5220f9d5acf864b58338c878" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.475825 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-tzhl4" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.575539 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr"] Sep 29 10:52:48 crc kubenswrapper[4727]: E0929 10:52:48.576182 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99127a9e-ae57-42b7-9392-e3f3d7ab43ff" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.576210 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="99127a9e-ae57-42b7-9392-e3f3d7ab43ff" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.576430 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="99127a9e-ae57-42b7-9392-e3f3d7ab43ff" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.577508 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.582005 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.582115 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.582599 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.582621 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.586653 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr"] Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.616864 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55kth\" (UniqueName: \"kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.617228 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.617423 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.719883 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.720040 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55kth\" (UniqueName: \"kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.720081 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.729247 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.731609 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.742199 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55kth\" (UniqueName: \"kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:48 crc kubenswrapper[4727]: I0929 10:52:48.906806 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:52:49 crc kubenswrapper[4727]: I0929 10:52:49.480554 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr"] Sep 29 10:52:50 crc kubenswrapper[4727]: I0929 10:52:50.506863 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" event={"ID":"d1b76565-6e67-46cf-9cc6-0f1b70d814a6","Type":"ContainerStarted","Data":"29ca7cbd69d9886f5419fd714671d61711b7bb72a6a03ddb4c35226254c7ce96"} Sep 29 10:52:50 crc kubenswrapper[4727]: I0929 10:52:50.507436 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" event={"ID":"d1b76565-6e67-46cf-9cc6-0f1b70d814a6","Type":"ContainerStarted","Data":"5c26bda70940a149d2e99af14e1267f1e95102adccb2c6d6e33a124018ec4fe4"} Sep 29 10:52:50 crc kubenswrapper[4727]: I0929 10:52:50.539508 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" podStartSLOduration=1.922420651 podStartE2EDuration="2.539480073s" podCreationTimestamp="2025-09-29 10:52:48 +0000 UTC" firstStartedPulling="2025-09-29 10:52:49.496422704 +0000 UTC m=+1839.669736066" lastFinishedPulling="2025-09-29 10:52:50.113482126 +0000 UTC m=+1840.286795488" observedRunningTime="2025-09-29 10:52:50.522315692 +0000 UTC m=+1840.695629074" watchObservedRunningTime="2025-09-29 10:52:50.539480073 +0000 UTC m=+1840.712793475" Sep 29 10:52:51 crc kubenswrapper[4727]: I0929 10:52:51.109637 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:52:51 crc kubenswrapper[4727]: I0929 10:52:51.540212 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f"} Sep 29 10:53:03 crc kubenswrapper[4727]: I0929 10:53:03.071230 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d7ljp"] Sep 29 10:53:03 crc kubenswrapper[4727]: I0929 10:53:03.086635 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d7ljp"] Sep 29 10:53:03 crc kubenswrapper[4727]: I0929 10:53:03.121205 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3866d914-1e2b-43e2-bf65-93d5e506fae2" path="/var/lib/kubelet/pods/3866d914-1e2b-43e2-bf65-93d5e506fae2/volumes" Sep 29 10:53:16 crc kubenswrapper[4727]: I0929 10:53:16.037289 4727 scope.go:117] "RemoveContainer" containerID="f6adc8347eccca98ece20adbd665742a43a470278d1c0f9dcb1f598e250ae4a9" Sep 29 10:53:16 crc kubenswrapper[4727]: I0929 10:53:16.084351 4727 scope.go:117] "RemoveContainer" containerID="fd0caed0286b93020f6a5945968a273ec5c0414a5b208da2d9f07a4d1ba2da6d" Sep 29 10:53:16 crc kubenswrapper[4727]: I0929 10:53:16.116915 4727 scope.go:117] "RemoveContainer" containerID="098ce89618430a91cc49dd15aef7f0be37b7a36fd22c333252cc8208db10684c" Sep 29 10:53:27 crc kubenswrapper[4727]: I0929 10:53:27.067570 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-dl7pr"] Sep 29 10:53:27 crc kubenswrapper[4727]: I0929 10:53:27.085047 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-dl7pr"] Sep 29 10:53:27 crc kubenswrapper[4727]: I0929 10:53:27.117842 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da99e317-fa23-4a56-a23d-1447b1ad3cb0" path="/var/lib/kubelet/pods/da99e317-fa23-4a56-a23d-1447b1ad3cb0/volumes" Sep 29 10:53:32 crc kubenswrapper[4727]: I0929 10:53:32.036426 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8n5tt"] Sep 29 10:53:32 crc kubenswrapper[4727]: I0929 10:53:32.047605 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8n5tt"] Sep 29 10:53:33 crc kubenswrapper[4727]: I0929 10:53:33.125802 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6027aeb9-0794-453e-85e3-9bf0e0c35d61" path="/var/lib/kubelet/pods/6027aeb9-0794-453e-85e3-9bf0e0c35d61/volumes" Sep 29 10:53:39 crc kubenswrapper[4727]: I0929 10:53:39.016330 4727 generic.go:334] "Generic (PLEG): container finished" podID="d1b76565-6e67-46cf-9cc6-0f1b70d814a6" containerID="29ca7cbd69d9886f5419fd714671d61711b7bb72a6a03ddb4c35226254c7ce96" exitCode=0 Sep 29 10:53:39 crc kubenswrapper[4727]: I0929 10:53:39.016455 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" event={"ID":"d1b76565-6e67-46cf-9cc6-0f1b70d814a6","Type":"ContainerDied","Data":"29ca7cbd69d9886f5419fd714671d61711b7bb72a6a03ddb4c35226254c7ce96"} Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.453349 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.511295 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory\") pod \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.511626 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55kth\" (UniqueName: \"kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth\") pod \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.511745 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key\") pod \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\" (UID: \"d1b76565-6e67-46cf-9cc6-0f1b70d814a6\") " Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.532573 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth" (OuterVolumeSpecName: "kube-api-access-55kth") pod "d1b76565-6e67-46cf-9cc6-0f1b70d814a6" (UID: "d1b76565-6e67-46cf-9cc6-0f1b70d814a6"). InnerVolumeSpecName "kube-api-access-55kth". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.538729 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory" (OuterVolumeSpecName: "inventory") pod "d1b76565-6e67-46cf-9cc6-0f1b70d814a6" (UID: "d1b76565-6e67-46cf-9cc6-0f1b70d814a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.540234 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d1b76565-6e67-46cf-9cc6-0f1b70d814a6" (UID: "d1b76565-6e67-46cf-9cc6-0f1b70d814a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.613642 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.613679 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55kth\" (UniqueName: \"kubernetes.io/projected/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-kube-api-access-55kth\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:40 crc kubenswrapper[4727]: I0929 10:53:40.613692 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1b76565-6e67-46cf-9cc6-0f1b70d814a6-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.035125 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" event={"ID":"d1b76565-6e67-46cf-9cc6-0f1b70d814a6","Type":"ContainerDied","Data":"5c26bda70940a149d2e99af14e1267f1e95102adccb2c6d6e33a124018ec4fe4"} Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.035617 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c26bda70940a149d2e99af14e1267f1e95102adccb2c6d6e33a124018ec4fe4" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.035189 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.127663 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vx2sv"] Sep 29 10:53:41 crc kubenswrapper[4727]: E0929 10:53:41.128105 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b76565-6e67-46cf-9cc6-0f1b70d814a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.128130 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b76565-6e67-46cf-9cc6-0f1b70d814a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.128389 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b76565-6e67-46cf-9cc6-0f1b70d814a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.129175 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.130381 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vx2sv"] Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.131388 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.131683 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.131843 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.132078 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.224376 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz86j\" (UniqueName: \"kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.224708 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.224847 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.326845 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz86j\" (UniqueName: \"kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.326948 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.326970 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.330721 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.330802 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.355225 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz86j\" (UniqueName: \"kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j\") pod \"ssh-known-hosts-edpm-deployment-vx2sv\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:41 crc kubenswrapper[4727]: I0929 10:53:41.447391 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:42 crc kubenswrapper[4727]: I0929 10:53:42.033831 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vx2sv"] Sep 29 10:53:42 crc kubenswrapper[4727]: W0929 10:53:42.037960 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28a41bdb_b145_481a_9835_e08ce95e2e17.slice/crio-4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502 WatchSource:0}: Error finding container 4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502: Status 404 returned error can't find the container with id 4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502 Sep 29 10:53:43 crc kubenswrapper[4727]: I0929 10:53:43.051954 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" event={"ID":"28a41bdb-b145-481a-9835-e08ce95e2e17","Type":"ContainerStarted","Data":"4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502"} Sep 29 10:53:44 crc kubenswrapper[4727]: I0929 10:53:44.061593 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" event={"ID":"28a41bdb-b145-481a-9835-e08ce95e2e17","Type":"ContainerStarted","Data":"541ad760bcc0fb3b88226dd014f9c2a780159dc5f7f098a83612b8cf500132df"} Sep 29 10:53:51 crc kubenswrapper[4727]: I0929 10:53:51.131006 4727 generic.go:334] "Generic (PLEG): container finished" podID="28a41bdb-b145-481a-9835-e08ce95e2e17" containerID="541ad760bcc0fb3b88226dd014f9c2a780159dc5f7f098a83612b8cf500132df" exitCode=0 Sep 29 10:53:51 crc kubenswrapper[4727]: I0929 10:53:51.131127 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" event={"ID":"28a41bdb-b145-481a-9835-e08ce95e2e17","Type":"ContainerDied","Data":"541ad760bcc0fb3b88226dd014f9c2a780159dc5f7f098a83612b8cf500132df"} Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.595857 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.672055 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam\") pod \"28a41bdb-b145-481a-9835-e08ce95e2e17\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.672203 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0\") pod \"28a41bdb-b145-481a-9835-e08ce95e2e17\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.672290 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz86j\" (UniqueName: \"kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j\") pod \"28a41bdb-b145-481a-9835-e08ce95e2e17\" (UID: \"28a41bdb-b145-481a-9835-e08ce95e2e17\") " Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.681351 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j" (OuterVolumeSpecName: "kube-api-access-mz86j") pod "28a41bdb-b145-481a-9835-e08ce95e2e17" (UID: "28a41bdb-b145-481a-9835-e08ce95e2e17"). InnerVolumeSpecName "kube-api-access-mz86j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.709787 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "28a41bdb-b145-481a-9835-e08ce95e2e17" (UID: "28a41bdb-b145-481a-9835-e08ce95e2e17"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.715892 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "28a41bdb-b145-481a-9835-e08ce95e2e17" (UID: "28a41bdb-b145-481a-9835-e08ce95e2e17"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.775187 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.775228 4727 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/28a41bdb-b145-481a-9835-e08ce95e2e17-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:52 crc kubenswrapper[4727]: I0929 10:53:52.775240 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz86j\" (UniqueName: \"kubernetes.io/projected/28a41bdb-b145-481a-9835-e08ce95e2e17-kube-api-access-mz86j\") on node \"crc\" DevicePath \"\"" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.151872 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" event={"ID":"28a41bdb-b145-481a-9835-e08ce95e2e17","Type":"ContainerDied","Data":"4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502"} Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.151945 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e4f9dfe8705985093d72691e2728c9e6ca6892f1a0bee1897d74b96e098f502" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.151980 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vx2sv" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.240947 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh"] Sep 29 10:53:53 crc kubenswrapper[4727]: E0929 10:53:53.242667 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a41bdb-b145-481a-9835-e08ce95e2e17" containerName="ssh-known-hosts-edpm-deployment" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.242734 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a41bdb-b145-481a-9835-e08ce95e2e17" containerName="ssh-known-hosts-edpm-deployment" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.243344 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a41bdb-b145-481a-9835-e08ce95e2e17" containerName="ssh-known-hosts-edpm-deployment" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.244770 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.254511 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh"] Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.267268 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.267581 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.268375 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.269712 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.392105 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.392194 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkhx\" (UniqueName: \"kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.392267 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.494434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.494509 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkhx\" (UniqueName: \"kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.494530 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.501556 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.502117 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.515079 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkhx\" (UniqueName: \"kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tfmmh\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:53 crc kubenswrapper[4727]: I0929 10:53:53.599115 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:53:54 crc kubenswrapper[4727]: I0929 10:53:54.160397 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh"] Sep 29 10:53:55 crc kubenswrapper[4727]: I0929 10:53:55.170699 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" event={"ID":"bc9712d6-5bbe-4749-acc0-a252a76a0ada","Type":"ContainerStarted","Data":"359a4f391035353384c30b89a2627f0fa8b7cd0064c9d18f48e8d454bd8a2ea5"} Sep 29 10:53:55 crc kubenswrapper[4727]: I0929 10:53:55.171421 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" event={"ID":"bc9712d6-5bbe-4749-acc0-a252a76a0ada","Type":"ContainerStarted","Data":"84a051a4173822d12676f324f227d7dba5083b07e2c2269e5b9de2519486016e"} Sep 29 10:53:55 crc kubenswrapper[4727]: I0929 10:53:55.196239 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" podStartSLOduration=1.620816926 podStartE2EDuration="2.196204516s" podCreationTimestamp="2025-09-29 10:53:53 +0000 UTC" firstStartedPulling="2025-09-29 10:53:54.16841954 +0000 UTC m=+1904.341732902" lastFinishedPulling="2025-09-29 10:53:54.74380712 +0000 UTC m=+1904.917120492" observedRunningTime="2025-09-29 10:53:55.192357817 +0000 UTC m=+1905.365671179" watchObservedRunningTime="2025-09-29 10:53:55.196204516 +0000 UTC m=+1905.369517878" Sep 29 10:54:03 crc kubenswrapper[4727]: I0929 10:54:03.263666 4727 generic.go:334] "Generic (PLEG): container finished" podID="bc9712d6-5bbe-4749-acc0-a252a76a0ada" containerID="359a4f391035353384c30b89a2627f0fa8b7cd0064c9d18f48e8d454bd8a2ea5" exitCode=0 Sep 29 10:54:03 crc kubenswrapper[4727]: I0929 10:54:03.263871 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" event={"ID":"bc9712d6-5bbe-4749-acc0-a252a76a0ada","Type":"ContainerDied","Data":"359a4f391035353384c30b89a2627f0fa8b7cd0064c9d18f48e8d454bd8a2ea5"} Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.654995 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.672437 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtkhx\" (UniqueName: \"kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx\") pod \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.672520 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory\") pod \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.672662 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key\") pod \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\" (UID: \"bc9712d6-5bbe-4749-acc0-a252a76a0ada\") " Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.688413 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx" (OuterVolumeSpecName: "kube-api-access-wtkhx") pod "bc9712d6-5bbe-4749-acc0-a252a76a0ada" (UID: "bc9712d6-5bbe-4749-acc0-a252a76a0ada"). InnerVolumeSpecName "kube-api-access-wtkhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.710477 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory" (OuterVolumeSpecName: "inventory") pod "bc9712d6-5bbe-4749-acc0-a252a76a0ada" (UID: "bc9712d6-5bbe-4749-acc0-a252a76a0ada"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.732502 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc9712d6-5bbe-4749-acc0-a252a76a0ada" (UID: "bc9712d6-5bbe-4749-acc0-a252a76a0ada"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.774684 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtkhx\" (UniqueName: \"kubernetes.io/projected/bc9712d6-5bbe-4749-acc0-a252a76a0ada-kube-api-access-wtkhx\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.774723 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:04 crc kubenswrapper[4727]: I0929 10:54:04.774735 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc9712d6-5bbe-4749-acc0-a252a76a0ada-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.282316 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" event={"ID":"bc9712d6-5bbe-4749-acc0-a252a76a0ada","Type":"ContainerDied","Data":"84a051a4173822d12676f324f227d7dba5083b07e2c2269e5b9de2519486016e"} Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.282372 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tfmmh" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.282397 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84a051a4173822d12676f324f227d7dba5083b07e2c2269e5b9de2519486016e" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.370567 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx"] Sep 29 10:54:05 crc kubenswrapper[4727]: E0929 10:54:05.371021 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9712d6-5bbe-4749-acc0-a252a76a0ada" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.371035 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9712d6-5bbe-4749-acc0-a252a76a0ada" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.371257 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc9712d6-5bbe-4749-acc0-a252a76a0ada" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.372038 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.374936 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.374991 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.375158 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.375351 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.385032 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx"] Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.387286 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.387521 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.387738 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8mkr\" (UniqueName: \"kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.490162 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.490240 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.490310 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8mkr\" (UniqueName: \"kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.494779 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.500956 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.513907 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8mkr\" (UniqueName: \"kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:05 crc kubenswrapper[4727]: I0929 10:54:05.691844 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:06 crc kubenswrapper[4727]: I0929 10:54:06.250087 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx"] Sep 29 10:54:06 crc kubenswrapper[4727]: W0929 10:54:06.255095 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6b1b4ed_2c0b_41f3_a542_e6fefaccc416.slice/crio-feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695 WatchSource:0}: Error finding container feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695: Status 404 returned error can't find the container with id feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695 Sep 29 10:54:06 crc kubenswrapper[4727]: I0929 10:54:06.295307 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" event={"ID":"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416","Type":"ContainerStarted","Data":"feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695"} Sep 29 10:54:07 crc kubenswrapper[4727]: I0929 10:54:07.306243 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" event={"ID":"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416","Type":"ContainerStarted","Data":"4e3205df551d0a84b9f97472316ece21d555c1b48895bb0ec1b64a51a32dc590"} Sep 29 10:54:07 crc kubenswrapper[4727]: I0929 10:54:07.325945 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" podStartSLOduration=1.634475831 podStartE2EDuration="2.325923976s" podCreationTimestamp="2025-09-29 10:54:05 +0000 UTC" firstStartedPulling="2025-09-29 10:54:06.257659189 +0000 UTC m=+1916.430972551" lastFinishedPulling="2025-09-29 10:54:06.949107334 +0000 UTC m=+1917.122420696" observedRunningTime="2025-09-29 10:54:07.322399496 +0000 UTC m=+1917.495712858" watchObservedRunningTime="2025-09-29 10:54:07.325923976 +0000 UTC m=+1917.499237338" Sep 29 10:54:11 crc kubenswrapper[4727]: I0929 10:54:11.041817 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-m77vs"] Sep 29 10:54:11 crc kubenswrapper[4727]: I0929 10:54:11.053082 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-m77vs"] Sep 29 10:54:11 crc kubenswrapper[4727]: I0929 10:54:11.122959 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbd9580-fa48-48d9-bcf1-4009c21d31a2" path="/var/lib/kubelet/pods/5cbd9580-fa48-48d9-bcf1-4009c21d31a2/volumes" Sep 29 10:54:16 crc kubenswrapper[4727]: I0929 10:54:16.209501 4727 scope.go:117] "RemoveContainer" containerID="4cd2c1e0dcfb72c0d3f22f6f8d299f86231c1885b308943dd7207aea3df69e55" Sep 29 10:54:16 crc kubenswrapper[4727]: I0929 10:54:16.276593 4727 scope.go:117] "RemoveContainer" containerID="8003c89863fa6c45ea034d993d33632f1c50aeb576150ec5a450e490934ebcff" Sep 29 10:54:16 crc kubenswrapper[4727]: I0929 10:54:16.357639 4727 scope.go:117] "RemoveContainer" containerID="a34b5738bfed82f755ab116015123779cad6e88fe1eb7e97a94f7e491172bf58" Sep 29 10:54:16 crc kubenswrapper[4727]: I0929 10:54:16.396035 4727 generic.go:334] "Generic (PLEG): container finished" podID="b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" containerID="4e3205df551d0a84b9f97472316ece21d555c1b48895bb0ec1b64a51a32dc590" exitCode=0 Sep 29 10:54:16 crc kubenswrapper[4727]: I0929 10:54:16.396601 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" event={"ID":"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416","Type":"ContainerDied","Data":"4e3205df551d0a84b9f97472316ece21d555c1b48895bb0ec1b64a51a32dc590"} Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.871717 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.947535 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory\") pod \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.947611 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key\") pod \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.947794 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8mkr\" (UniqueName: \"kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr\") pod \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\" (UID: \"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416\") " Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.954738 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr" (OuterVolumeSpecName: "kube-api-access-r8mkr") pod "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" (UID: "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416"). InnerVolumeSpecName "kube-api-access-r8mkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.983149 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" (UID: "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:54:17 crc kubenswrapper[4727]: I0929 10:54:17.983204 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory" (OuterVolumeSpecName: "inventory") pod "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" (UID: "b6b1b4ed-2c0b-41f3-a542-e6fefaccc416"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.051629 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.051683 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.051705 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8mkr\" (UniqueName: \"kubernetes.io/projected/b6b1b4ed-2c0b-41f3-a542-e6fefaccc416-kube-api-access-r8mkr\") on node \"crc\" DevicePath \"\"" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.430050 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" event={"ID":"b6b1b4ed-2c0b-41f3-a542-e6fefaccc416","Type":"ContainerDied","Data":"feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695"} Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.430579 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feed4736625be6c6e7e0c7bf4d673b6e1a241f153431f37291587ae31549d695" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.430302 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.502244 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72"] Sep 29 10:54:18 crc kubenswrapper[4727]: E0929 10:54:18.502745 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.502770 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.502947 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b1b4ed-2c0b-41f3-a542-e6fefaccc416" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.503662 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.510966 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.512056 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.512083 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.512637 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.512660 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.512838 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.515857 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.516105 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.522805 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72"] Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.560623 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.560714 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.560827 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.560862 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.560945 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561000 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561048 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561133 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561210 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561304 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561362 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561503 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561563 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df84m\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.561593 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663670 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663727 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df84m\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663775 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663818 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663843 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663898 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663930 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663964 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.663992 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.664016 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.664065 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.664089 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.664168 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.664203 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.672681 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.672680 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.672884 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.673870 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.673973 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.674576 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.674706 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.675051 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.675209 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.675730 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.675820 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.676618 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.677658 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.685118 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df84m\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xqw72\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:18 crc kubenswrapper[4727]: I0929 10:54:18.828814 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:54:19 crc kubenswrapper[4727]: I0929 10:54:19.469309 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72"] Sep 29 10:54:19 crc kubenswrapper[4727]: I0929 10:54:19.494371 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:54:20 crc kubenswrapper[4727]: I0929 10:54:20.452882 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" event={"ID":"d3ec4e67-b454-43bd-ab50-0257dc2e517d","Type":"ContainerStarted","Data":"cee8d42620f549f6efd5d0cd038b23e43191c739dd63e9d3aff58da86786edd3"} Sep 29 10:54:20 crc kubenswrapper[4727]: I0929 10:54:20.453674 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" event={"ID":"d3ec4e67-b454-43bd-ab50-0257dc2e517d","Type":"ContainerStarted","Data":"8c14e1a9de71e03b811e61b9c920d2888615cf4fc827e440b9709c1982bf8f9c"} Sep 29 10:54:58 crc kubenswrapper[4727]: I0929 10:54:58.798779 4727 generic.go:334] "Generic (PLEG): container finished" podID="d3ec4e67-b454-43bd-ab50-0257dc2e517d" containerID="cee8d42620f549f6efd5d0cd038b23e43191c739dd63e9d3aff58da86786edd3" exitCode=0 Sep 29 10:54:58 crc kubenswrapper[4727]: I0929 10:54:58.798922 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" event={"ID":"d3ec4e67-b454-43bd-ab50-0257dc2e517d","Type":"ContainerDied","Data":"cee8d42620f549f6efd5d0cd038b23e43191c739dd63e9d3aff58da86786edd3"} Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.233827 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408104 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408614 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408653 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408766 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408841 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df84m\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408886 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408922 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408962 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.408995 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.409065 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.409116 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.409157 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.409239 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.409301 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle\") pod \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\" (UID: \"d3ec4e67-b454-43bd-ab50-0257dc2e517d\") " Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.417429 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.417425 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.417790 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.419719 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.419751 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.419743 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.419776 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m" (OuterVolumeSpecName: "kube-api-access-df84m") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "kube-api-access-df84m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.419969 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.422054 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.422311 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.422706 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.425070 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.447871 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory" (OuterVolumeSpecName: "inventory") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.448419 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3ec4e67-b454-43bd-ab50-0257dc2e517d" (UID: "d3ec4e67-b454-43bd-ab50-0257dc2e517d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511542 4727 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511587 4727 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511601 4727 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511611 4727 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511620 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df84m\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-kube-api-access-df84m\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511631 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511646 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511655 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511666 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511676 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511687 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d3ec4e67-b454-43bd-ab50-0257dc2e517d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511696 4727 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511705 4727 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.511715 4727 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ec4e67-b454-43bd-ab50-0257dc2e517d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.819490 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" event={"ID":"d3ec4e67-b454-43bd-ab50-0257dc2e517d","Type":"ContainerDied","Data":"8c14e1a9de71e03b811e61b9c920d2888615cf4fc827e440b9709c1982bf8f9c"} Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.819895 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c14e1a9de71e03b811e61b9c920d2888615cf4fc827e440b9709c1982bf8f9c" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.819554 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xqw72" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.924723 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq"] Sep 29 10:55:00 crc kubenswrapper[4727]: E0929 10:55:00.925127 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ec4e67-b454-43bd-ab50-0257dc2e517d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.925141 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ec4e67-b454-43bd-ab50-0257dc2e517d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.925329 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ec4e67-b454-43bd-ab50-0257dc2e517d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.926033 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.929497 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.929744 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.929967 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.930141 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.930518 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:55:00 crc kubenswrapper[4727]: I0929 10:55:00.940843 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq"] Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.020154 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tqv4\" (UniqueName: \"kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.020208 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.020273 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.020312 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.020362 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.121772 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.121853 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.121896 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.121936 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.121989 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tqv4\" (UniqueName: \"kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.123062 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.126841 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.130140 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.146229 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.149225 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tqv4\" (UniqueName: \"kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9gxq\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.251778 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.765911 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq"] Sep 29 10:55:01 crc kubenswrapper[4727]: I0929 10:55:01.828694 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" event={"ID":"1dc57566-f99c-4041-a617-199b93c4e4eb","Type":"ContainerStarted","Data":"716b3d07ac639b1900876dbd622f44cd3a62d9947580a43b67c7e524630f543d"} Sep 29 10:55:02 crc kubenswrapper[4727]: I0929 10:55:02.864853 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" event={"ID":"1dc57566-f99c-4041-a617-199b93c4e4eb","Type":"ContainerStarted","Data":"4f1445693a5e82689a501d986faa509a1f3c46f7fb934bc74ce771a719b7ebe1"} Sep 29 10:55:02 crc kubenswrapper[4727]: I0929 10:55:02.885749 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" podStartSLOduration=2.436953278 podStartE2EDuration="2.885726171s" podCreationTimestamp="2025-09-29 10:55:00 +0000 UTC" firstStartedPulling="2025-09-29 10:55:01.781686264 +0000 UTC m=+1971.954999626" lastFinishedPulling="2025-09-29 10:55:02.230459157 +0000 UTC m=+1972.403772519" observedRunningTime="2025-09-29 10:55:02.884784707 +0000 UTC m=+1973.058098069" watchObservedRunningTime="2025-09-29 10:55:02.885726171 +0000 UTC m=+1973.059039543" Sep 29 10:55:19 crc kubenswrapper[4727]: I0929 10:55:19.246770 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:55:19 crc kubenswrapper[4727]: I0929 10:55:19.247556 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:55:49 crc kubenswrapper[4727]: I0929 10:55:49.246968 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:55:49 crc kubenswrapper[4727]: I0929 10:55:49.247967 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:56:01 crc kubenswrapper[4727]: I0929 10:56:01.429100 4727 generic.go:334] "Generic (PLEG): container finished" podID="1dc57566-f99c-4041-a617-199b93c4e4eb" containerID="4f1445693a5e82689a501d986faa509a1f3c46f7fb934bc74ce771a719b7ebe1" exitCode=0 Sep 29 10:56:01 crc kubenswrapper[4727]: I0929 10:56:01.429198 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" event={"ID":"1dc57566-f99c-4041-a617-199b93c4e4eb","Type":"ContainerDied","Data":"4f1445693a5e82689a501d986faa509a1f3c46f7fb934bc74ce771a719b7ebe1"} Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.824959 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.852431 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle\") pod \"1dc57566-f99c-4041-a617-199b93c4e4eb\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.852484 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0\") pod \"1dc57566-f99c-4041-a617-199b93c4e4eb\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.852573 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory\") pod \"1dc57566-f99c-4041-a617-199b93c4e4eb\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.852659 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key\") pod \"1dc57566-f99c-4041-a617-199b93c4e4eb\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.852684 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tqv4\" (UniqueName: \"kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4\") pod \"1dc57566-f99c-4041-a617-199b93c4e4eb\" (UID: \"1dc57566-f99c-4041-a617-199b93c4e4eb\") " Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.863138 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4" (OuterVolumeSpecName: "kube-api-access-7tqv4") pod "1dc57566-f99c-4041-a617-199b93c4e4eb" (UID: "1dc57566-f99c-4041-a617-199b93c4e4eb"). InnerVolumeSpecName "kube-api-access-7tqv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.863365 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1dc57566-f99c-4041-a617-199b93c4e4eb" (UID: "1dc57566-f99c-4041-a617-199b93c4e4eb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.883684 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1dc57566-f99c-4041-a617-199b93c4e4eb" (UID: "1dc57566-f99c-4041-a617-199b93c4e4eb"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.885492 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory" (OuterVolumeSpecName: "inventory") pod "1dc57566-f99c-4041-a617-199b93c4e4eb" (UID: "1dc57566-f99c-4041-a617-199b93c4e4eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.897067 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1dc57566-f99c-4041-a617-199b93c4e4eb" (UID: "1dc57566-f99c-4041-a617-199b93c4e4eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.954271 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.954302 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.954313 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tqv4\" (UniqueName: \"kubernetes.io/projected/1dc57566-f99c-4041-a617-199b93c4e4eb-kube-api-access-7tqv4\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.954322 4727 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc57566-f99c-4041-a617-199b93c4e4eb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:02 crc kubenswrapper[4727]: I0929 10:56:02.954346 4727 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1dc57566-f99c-4041-a617-199b93c4e4eb-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.453204 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" event={"ID":"1dc57566-f99c-4041-a617-199b93c4e4eb","Type":"ContainerDied","Data":"716b3d07ac639b1900876dbd622f44cd3a62d9947580a43b67c7e524630f543d"} Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.453517 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="716b3d07ac639b1900876dbd622f44cd3a62d9947580a43b67c7e524630f543d" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.453285 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9gxq" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.563115 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t"] Sep 29 10:56:03 crc kubenswrapper[4727]: E0929 10:56:03.563728 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc57566-f99c-4041-a617-199b93c4e4eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.563751 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc57566-f99c-4041-a617-199b93c4e4eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.564022 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc57566-f99c-4041-a617-199b93c4e4eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.565136 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.572020 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t"] Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580091 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580249 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580312 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580373 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580420 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.580528 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.670956 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.671020 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.671080 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.671156 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.671197 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9ql5\" (UniqueName: \"kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.671236 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.773291 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9ql5\" (UniqueName: \"kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.773391 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.773498 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.773541 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.773579 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.775618 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.781520 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.782023 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.782652 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.783820 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.789470 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.794766 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9ql5\" (UniqueName: \"kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:03 crc kubenswrapper[4727]: I0929 10:56:03.902875 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:04 crc kubenswrapper[4727]: I0929 10:56:04.484107 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t"] Sep 29 10:56:05 crc kubenswrapper[4727]: I0929 10:56:05.471715 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" event={"ID":"4bd4cabe-bcac-4332-b7be-326693d4bc95","Type":"ContainerStarted","Data":"0d0b740d839a65d6ea148fdf6a78fcb3e6df14b7b4627ae22cd559e692c8f1b5"} Sep 29 10:56:05 crc kubenswrapper[4727]: I0929 10:56:05.471940 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" event={"ID":"4bd4cabe-bcac-4332-b7be-326693d4bc95","Type":"ContainerStarted","Data":"ad945240ffb1a0b5dab74f997888ec6ebf7b5e87d5abacae8a9bca705a599250"} Sep 29 10:56:05 crc kubenswrapper[4727]: I0929 10:56:05.495809 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" podStartSLOduration=1.907997093 podStartE2EDuration="2.495774648s" podCreationTimestamp="2025-09-29 10:56:03 +0000 UTC" firstStartedPulling="2025-09-29 10:56:04.489164317 +0000 UTC m=+2034.662477679" lastFinishedPulling="2025-09-29 10:56:05.076941872 +0000 UTC m=+2035.250255234" observedRunningTime="2025-09-29 10:56:05.487993689 +0000 UTC m=+2035.661307051" watchObservedRunningTime="2025-09-29 10:56:05.495774648 +0000 UTC m=+2035.669088010" Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.246904 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.247579 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.247635 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.248729 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.248790 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f" gracePeriod=600 Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.603938 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f" exitCode=0 Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.604011 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f"} Sep 29 10:56:19 crc kubenswrapper[4727]: I0929 10:56:19.604043 4727 scope.go:117] "RemoveContainer" containerID="d92e93f260847805aefcc722ad198a6670ab4072eb395039123d2c90100938a4" Sep 29 10:56:20 crc kubenswrapper[4727]: I0929 10:56:20.614270 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3"} Sep 29 10:56:50 crc kubenswrapper[4727]: I0929 10:56:50.905701 4727 generic.go:334] "Generic (PLEG): container finished" podID="4bd4cabe-bcac-4332-b7be-326693d4bc95" containerID="0d0b740d839a65d6ea148fdf6a78fcb3e6df14b7b4627ae22cd559e692c8f1b5" exitCode=0 Sep 29 10:56:50 crc kubenswrapper[4727]: I0929 10:56:50.905770 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" event={"ID":"4bd4cabe-bcac-4332-b7be-326693d4bc95","Type":"ContainerDied","Data":"0d0b740d839a65d6ea148fdf6a78fcb3e6df14b7b4627ae22cd559e692c8f1b5"} Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.337579 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.414294 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9ql5\" (UniqueName: \"kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.414392 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.414476 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.414502 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.415126 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.415151 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory\") pod \"4bd4cabe-bcac-4332-b7be-326693d4bc95\" (UID: \"4bd4cabe-bcac-4332-b7be-326693d4bc95\") " Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.420207 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.420815 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5" (OuterVolumeSpecName: "kube-api-access-k9ql5") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "kube-api-access-k9ql5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.443561 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory" (OuterVolumeSpecName: "inventory") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.446983 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.447393 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.451504 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4bd4cabe-bcac-4332-b7be-326693d4bc95" (UID: "4bd4cabe-bcac-4332-b7be-326693d4bc95"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517327 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9ql5\" (UniqueName: \"kubernetes.io/projected/4bd4cabe-bcac-4332-b7be-326693d4bc95-kube-api-access-k9ql5\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517391 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517406 4727 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517423 4727 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517436 4727 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.517448 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd4cabe-bcac-4332-b7be-326693d4bc95-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.930506 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" event={"ID":"4bd4cabe-bcac-4332-b7be-326693d4bc95","Type":"ContainerDied","Data":"ad945240ffb1a0b5dab74f997888ec6ebf7b5e87d5abacae8a9bca705a599250"} Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.930842 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad945240ffb1a0b5dab74f997888ec6ebf7b5e87d5abacae8a9bca705a599250" Sep 29 10:56:52 crc kubenswrapper[4727]: I0929 10:56:52.930565 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.014891 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj"] Sep 29 10:56:53 crc kubenswrapper[4727]: E0929 10:56:53.015401 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bd4cabe-bcac-4332-b7be-326693d4bc95" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.015430 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bd4cabe-bcac-4332-b7be-326693d4bc95" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.015684 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bd4cabe-bcac-4332-b7be-326693d4bc95" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.016460 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.023133 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj"] Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.023752 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.024100 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.024451 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.025536 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.029994 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.129835 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.129961 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgsmh\" (UniqueName: \"kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.130066 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.130102 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.130135 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.232434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.232485 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgsmh\" (UniqueName: \"kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.232584 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.232615 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.232662 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.237069 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.238551 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.239397 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.241586 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.251078 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgsmh\" (UniqueName: \"kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xskrj\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.335233 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.865692 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj"] Sep 29 10:56:53 crc kubenswrapper[4727]: I0929 10:56:53.939151 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" event={"ID":"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea","Type":"ContainerStarted","Data":"92afad48873d575d1e79bf370cfd0b3ceb2917bd4755269d6361c3f43141d4b9"} Sep 29 10:56:54 crc kubenswrapper[4727]: I0929 10:56:54.954916 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" event={"ID":"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea","Type":"ContainerStarted","Data":"96919151de6ae9292563c33960fb55b9580e09a07edb031a7b8fefcaaf7c68ee"} Sep 29 10:56:54 crc kubenswrapper[4727]: I0929 10:56:54.986327 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" podStartSLOduration=2.338531692 podStartE2EDuration="2.98630097s" podCreationTimestamp="2025-09-29 10:56:52 +0000 UTC" firstStartedPulling="2025-09-29 10:56:53.874934092 +0000 UTC m=+2084.048247454" lastFinishedPulling="2025-09-29 10:56:54.52270337 +0000 UTC m=+2084.696016732" observedRunningTime="2025-09-29 10:56:54.979593809 +0000 UTC m=+2085.152907171" watchObservedRunningTime="2025-09-29 10:56:54.98630097 +0000 UTC m=+2085.159614332" Sep 29 10:57:09 crc kubenswrapper[4727]: I0929 10:57:09.968735 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9s47w"] Sep 29 10:57:09 crc kubenswrapper[4727]: I0929 10:57:09.976482 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.008013 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9s47w"] Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.015060 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-utilities\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.015166 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t995\" (UniqueName: \"kubernetes.io/projected/b161c1a4-0830-45fa-943b-579070b738ff-kube-api-access-7t995\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.015315 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-catalog-content\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.117577 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-utilities\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.117668 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t995\" (UniqueName: \"kubernetes.io/projected/b161c1a4-0830-45fa-943b-579070b738ff-kube-api-access-7t995\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.117731 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-catalog-content\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.118220 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-catalog-content\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.118622 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b161c1a4-0830-45fa-943b-579070b738ff-utilities\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.140680 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t995\" (UniqueName: \"kubernetes.io/projected/b161c1a4-0830-45fa-943b-579070b738ff-kube-api-access-7t995\") pod \"redhat-operators-9s47w\" (UID: \"b161c1a4-0830-45fa-943b-579070b738ff\") " pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.333178 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:10 crc kubenswrapper[4727]: I0929 10:57:10.827053 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9s47w"] Sep 29 10:57:11 crc kubenswrapper[4727]: I0929 10:57:11.096009 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerStarted","Data":"035f32c43a6942a7bdefa92fbb124ae7e479f79bf0231bc4a9c4a67d01e14b77"} Sep 29 10:57:11 crc kubenswrapper[4727]: I0929 10:57:11.096057 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerStarted","Data":"754d00ec7850576cabe41f56b2827b072a2d476b5472e6e4a4122c2192d67fa9"} Sep 29 10:57:12 crc kubenswrapper[4727]: I0929 10:57:12.111255 4727 generic.go:334] "Generic (PLEG): container finished" podID="b161c1a4-0830-45fa-943b-579070b738ff" containerID="035f32c43a6942a7bdefa92fbb124ae7e479f79bf0231bc4a9c4a67d01e14b77" exitCode=0 Sep 29 10:57:12 crc kubenswrapper[4727]: I0929 10:57:12.111381 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerDied","Data":"035f32c43a6942a7bdefa92fbb124ae7e479f79bf0231bc4a9c4a67d01e14b77"} Sep 29 10:57:18 crc kubenswrapper[4727]: I0929 10:57:18.181397 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerStarted","Data":"b3f797b79cf171f809d87ace78d596551f7d221c5d9abb3592e66dcaf25f5187"} Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.201427 4727 generic.go:334] "Generic (PLEG): container finished" podID="b161c1a4-0830-45fa-943b-579070b738ff" containerID="b3f797b79cf171f809d87ace78d596551f7d221c5d9abb3592e66dcaf25f5187" exitCode=0 Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.201557 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerDied","Data":"b3f797b79cf171f809d87ace78d596551f7d221c5d9abb3592e66dcaf25f5187"} Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.777370 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.780636 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.788733 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.915850 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.916751 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drvgs\" (UniqueName: \"kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:20 crc kubenswrapper[4727]: I0929 10:57:20.916996 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.019226 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.019573 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.019682 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drvgs\" (UniqueName: \"kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.019980 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.020528 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.049646 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drvgs\" (UniqueName: \"kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs\") pod \"redhat-marketplace-4dt85\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.150559 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:21 crc kubenswrapper[4727]: I0929 10:57:21.639256 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:21 crc kubenswrapper[4727]: W0929 10:57:21.650004 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99ca2de8_9cd7_4ff0_b523_08ccb148a297.slice/crio-5960eb6ae38a9b2faad02150fbdf3ba270548a8447a52be7ccf0398eda959f49 WatchSource:0}: Error finding container 5960eb6ae38a9b2faad02150fbdf3ba270548a8447a52be7ccf0398eda959f49: Status 404 returned error can't find the container with id 5960eb6ae38a9b2faad02150fbdf3ba270548a8447a52be7ccf0398eda959f49 Sep 29 10:57:22 crc kubenswrapper[4727]: I0929 10:57:22.221057 4727 generic.go:334] "Generic (PLEG): container finished" podID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerID="3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7" exitCode=0 Sep 29 10:57:22 crc kubenswrapper[4727]: I0929 10:57:22.221577 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerDied","Data":"3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7"} Sep 29 10:57:22 crc kubenswrapper[4727]: I0929 10:57:22.221605 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerStarted","Data":"5960eb6ae38a9b2faad02150fbdf3ba270548a8447a52be7ccf0398eda959f49"} Sep 29 10:57:22 crc kubenswrapper[4727]: I0929 10:57:22.224700 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s47w" event={"ID":"b161c1a4-0830-45fa-943b-579070b738ff","Type":"ContainerStarted","Data":"047fe707a6153d4130e60f4dbcf19b62e4395b6c4d905bfc6934ce2e88e3d41c"} Sep 29 10:57:22 crc kubenswrapper[4727]: I0929 10:57:22.263120 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9s47w" podStartSLOduration=2.817025602 podStartE2EDuration="13.263104504s" podCreationTimestamp="2025-09-29 10:57:09 +0000 UTC" firstStartedPulling="2025-09-29 10:57:11.098368172 +0000 UTC m=+2101.271681534" lastFinishedPulling="2025-09-29 10:57:21.544447064 +0000 UTC m=+2111.717760436" observedRunningTime="2025-09-29 10:57:22.256872315 +0000 UTC m=+2112.430185677" watchObservedRunningTime="2025-09-29 10:57:22.263104504 +0000 UTC m=+2112.436417866" Sep 29 10:57:23 crc kubenswrapper[4727]: I0929 10:57:23.237701 4727 generic.go:334] "Generic (PLEG): container finished" podID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerID="1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200" exitCode=0 Sep 29 10:57:23 crc kubenswrapper[4727]: I0929 10:57:23.237824 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerDied","Data":"1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200"} Sep 29 10:57:27 crc kubenswrapper[4727]: I0929 10:57:27.291910 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerStarted","Data":"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8"} Sep 29 10:57:27 crc kubenswrapper[4727]: I0929 10:57:27.322773 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4dt85" podStartSLOduration=5.857611323 podStartE2EDuration="7.322756734s" podCreationTimestamp="2025-09-29 10:57:20 +0000 UTC" firstStartedPulling="2025-09-29 10:57:22.224537678 +0000 UTC m=+2112.397851040" lastFinishedPulling="2025-09-29 10:57:23.689683079 +0000 UTC m=+2113.862996451" observedRunningTime="2025-09-29 10:57:27.32062651 +0000 UTC m=+2117.493939882" watchObservedRunningTime="2025-09-29 10:57:27.322756734 +0000 UTC m=+2117.496070096" Sep 29 10:57:30 crc kubenswrapper[4727]: I0929 10:57:30.334221 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:30 crc kubenswrapper[4727]: I0929 10:57:30.334521 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.152266 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.152698 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.201241 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.370068 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.378003 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9s47w" podUID="b161c1a4-0830-45fa-943b-579070b738ff" containerName="registry-server" probeResult="failure" output=< Sep 29 10:57:31 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 10:57:31 crc kubenswrapper[4727]: > Sep 29 10:57:31 crc kubenswrapper[4727]: I0929 10:57:31.441258 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.344757 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4dt85" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="registry-server" containerID="cri-o://d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8" gracePeriod=2 Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.748041 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.798425 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drvgs\" (UniqueName: \"kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs\") pod \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.798562 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content\") pod \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.798808 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities\") pod \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\" (UID: \"99ca2de8-9cd7-4ff0-b523-08ccb148a297\") " Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.799616 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities" (OuterVolumeSpecName: "utilities") pod "99ca2de8-9cd7-4ff0-b523-08ccb148a297" (UID: "99ca2de8-9cd7-4ff0-b523-08ccb148a297"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.807709 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs" (OuterVolumeSpecName: "kube-api-access-drvgs") pod "99ca2de8-9cd7-4ff0-b523-08ccb148a297" (UID: "99ca2de8-9cd7-4ff0-b523-08ccb148a297"). InnerVolumeSpecName "kube-api-access-drvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.821138 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99ca2de8-9cd7-4ff0-b523-08ccb148a297" (UID: "99ca2de8-9cd7-4ff0-b523-08ccb148a297"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.900640 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drvgs\" (UniqueName: \"kubernetes.io/projected/99ca2de8-9cd7-4ff0-b523-08ccb148a297-kube-api-access-drvgs\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.900681 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:33 crc kubenswrapper[4727]: I0929 10:57:33.900691 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ca2de8-9cd7-4ff0-b523-08ccb148a297-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.355723 4727 generic.go:334] "Generic (PLEG): container finished" podID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerID="d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8" exitCode=0 Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.355806 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dt85" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.355816 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerDied","Data":"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8"} Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.357097 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dt85" event={"ID":"99ca2de8-9cd7-4ff0-b523-08ccb148a297","Type":"ContainerDied","Data":"5960eb6ae38a9b2faad02150fbdf3ba270548a8447a52be7ccf0398eda959f49"} Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.357121 4727 scope.go:117] "RemoveContainer" containerID="d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.378504 4727 scope.go:117] "RemoveContainer" containerID="1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.391345 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.400013 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dt85"] Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.415615 4727 scope.go:117] "RemoveContainer" containerID="3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.450956 4727 scope.go:117] "RemoveContainer" containerID="d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8" Sep 29 10:57:34 crc kubenswrapper[4727]: E0929 10:57:34.451572 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8\": container with ID starting with d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8 not found: ID does not exist" containerID="d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.451615 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8"} err="failed to get container status \"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8\": rpc error: code = NotFound desc = could not find container \"d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8\": container with ID starting with d6c2faced4d1e5ec817a04c831a7856100158ddf9f62d3a9abe5a760a53087b8 not found: ID does not exist" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.451656 4727 scope.go:117] "RemoveContainer" containerID="1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200" Sep 29 10:57:34 crc kubenswrapper[4727]: E0929 10:57:34.452031 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200\": container with ID starting with 1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200 not found: ID does not exist" containerID="1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.452062 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200"} err="failed to get container status \"1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200\": rpc error: code = NotFound desc = could not find container \"1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200\": container with ID starting with 1e8208103b448c169a30a7177e5de75dee3cb473b10eac23d12b9933a708a200 not found: ID does not exist" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.452080 4727 scope.go:117] "RemoveContainer" containerID="3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7" Sep 29 10:57:34 crc kubenswrapper[4727]: E0929 10:57:34.452990 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7\": container with ID starting with 3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7 not found: ID does not exist" containerID="3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7" Sep 29 10:57:34 crc kubenswrapper[4727]: I0929 10:57:34.453024 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7"} err="failed to get container status \"3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7\": rpc error: code = NotFound desc = could not find container \"3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7\": container with ID starting with 3a17f3937b72dc2a2c91d4b6e5242079f63ef9b4209389fa0191d81055ad38a7 not found: ID does not exist" Sep 29 10:57:35 crc kubenswrapper[4727]: I0929 10:57:35.123249 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" path="/var/lib/kubelet/pods/99ca2de8-9cd7-4ff0-b523-08ccb148a297/volumes" Sep 29 10:57:41 crc kubenswrapper[4727]: I0929 10:57:41.379527 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9s47w" podUID="b161c1a4-0830-45fa-943b-579070b738ff" containerName="registry-server" probeResult="failure" output=< Sep 29 10:57:41 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 10:57:41 crc kubenswrapper[4727]: > Sep 29 10:57:50 crc kubenswrapper[4727]: I0929 10:57:50.379948 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:50 crc kubenswrapper[4727]: I0929 10:57:50.431866 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9s47w" Sep 29 10:57:50 crc kubenswrapper[4727]: I0929 10:57:50.497219 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9s47w"] Sep 29 10:57:50 crc kubenswrapper[4727]: I0929 10:57:50.614758 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:57:50 crc kubenswrapper[4727]: I0929 10:57:50.615023 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9zv28" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="registry-server" containerID="cri-o://70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280" gracePeriod=2 Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.071826 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.212533 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44hzs\" (UniqueName: \"kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs\") pod \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.212609 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities\") pod \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.212697 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content\") pod \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\" (UID: \"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c\") " Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.214603 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities" (OuterVolumeSpecName: "utilities") pod "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" (UID: "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.250201 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs" (OuterVolumeSpecName: "kube-api-access-44hzs") pod "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" (UID: "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c"). InnerVolumeSpecName "kube-api-access-44hzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.304061 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" (UID: "eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.315079 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.315121 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44hzs\" (UniqueName: \"kubernetes.io/projected/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-kube-api-access-44hzs\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.315153 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.538395 4727 generic.go:334] "Generic (PLEG): container finished" podID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerID="70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280" exitCode=0 Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.539425 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zv28" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.542036 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerDied","Data":"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280"} Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.542182 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zv28" event={"ID":"eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c","Type":"ContainerDied","Data":"c11e45aced53f8243a5b7cab9b56974627cd78c69120cb08f69068b7284db750"} Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.542263 4727 scope.go:117] "RemoveContainer" containerID="70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.565985 4727 scope.go:117] "RemoveContainer" containerID="03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.582390 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.591651 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9zv28"] Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.600268 4727 scope.go:117] "RemoveContainer" containerID="f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.643061 4727 scope.go:117] "RemoveContainer" containerID="70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280" Sep 29 10:57:51 crc kubenswrapper[4727]: E0929 10:57:51.643530 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280\": container with ID starting with 70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280 not found: ID does not exist" containerID="70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.643556 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280"} err="failed to get container status \"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280\": rpc error: code = NotFound desc = could not find container \"70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280\": container with ID starting with 70901636a7400a1e607332d9e1038ad8286dfddfb6da6b1653fdc4086911e280 not found: ID does not exist" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.643575 4727 scope.go:117] "RemoveContainer" containerID="03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3" Sep 29 10:57:51 crc kubenswrapper[4727]: E0929 10:57:51.643908 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3\": container with ID starting with 03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3 not found: ID does not exist" containerID="03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.643939 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3"} err="failed to get container status \"03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3\": rpc error: code = NotFound desc = could not find container \"03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3\": container with ID starting with 03c09dcc38369d06e1c43b5aed0429c3dbd45ec16bb219dfb9e40376cdefb5a3 not found: ID does not exist" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.644000 4727 scope.go:117] "RemoveContainer" containerID="f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d" Sep 29 10:57:51 crc kubenswrapper[4727]: E0929 10:57:51.644405 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d\": container with ID starting with f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d not found: ID does not exist" containerID="f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d" Sep 29 10:57:51 crc kubenswrapper[4727]: I0929 10:57:51.644434 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d"} err="failed to get container status \"f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d\": rpc error: code = NotFound desc = could not find container \"f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d\": container with ID starting with f219d698989fff7f79d6e3225a1fe4442f41a568b3716913ffac16956300e80d not found: ID does not exist" Sep 29 10:57:53 crc kubenswrapper[4727]: I0929 10:57:53.125641 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" path="/var/lib/kubelet/pods/eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c/volumes" Sep 29 10:58:19 crc kubenswrapper[4727]: I0929 10:58:19.246964 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:58:19 crc kubenswrapper[4727]: I0929 10:58:19.247477 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:58:49 crc kubenswrapper[4727]: I0929 10:58:49.247323 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:58:49 crc kubenswrapper[4727]: I0929 10:58:49.247976 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.247036 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.248795 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.248931 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.249749 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.249888 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" gracePeriod=600 Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.380415 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.477755 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478508 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="extract-content" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478528 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="extract-content" Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478549 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="extract-content" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478557 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="extract-content" Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478570 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="extract-utilities" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478577 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="extract-utilities" Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478605 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478611 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478619 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478625 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: E0929 10:59:19.478635 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="extract-utilities" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478640 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="extract-utilities" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478811 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="eddd788b-0bbb-4ffc-a3de-0dbfbb39e21c" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.478836 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ca2de8-9cd7-4ff0-b523-08ccb148a297" containerName="registry-server" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.481306 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.491424 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.624732 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.625029 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgqg\" (UniqueName: \"kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.625287 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.727099 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.727145 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgqg\" (UniqueName: \"kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.727289 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.727819 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.728132 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.762294 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgqg\" (UniqueName: \"kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg\") pod \"certified-operators-8nq2l\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:19 crc kubenswrapper[4727]: I0929 10:59:19.842144 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:20 crc kubenswrapper[4727]: I0929 10:59:20.364104 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" exitCode=0 Sep 29 10:59:20 crc kubenswrapper[4727]: I0929 10:59:20.364310 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3"} Sep 29 10:59:20 crc kubenswrapper[4727]: I0929 10:59:20.364392 4727 scope.go:117] "RemoveContainer" containerID="263e4a6f6863e5bc23a09da9dcd99195693269fde7804eb5d99bb72c9472f21f" Sep 29 10:59:20 crc kubenswrapper[4727]: I0929 10:59:20.364976 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 10:59:20 crc kubenswrapper[4727]: E0929 10:59:20.365373 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:59:20 crc kubenswrapper[4727]: I0929 10:59:20.370459 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:21 crc kubenswrapper[4727]: I0929 10:59:21.377916 4727 generic.go:334] "Generic (PLEG): container finished" podID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerID="6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af" exitCode=0 Sep 29 10:59:21 crc kubenswrapper[4727]: I0929 10:59:21.378063 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerDied","Data":"6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af"} Sep 29 10:59:21 crc kubenswrapper[4727]: I0929 10:59:21.378524 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerStarted","Data":"b50336cbe1f83c469289526b108902a5542ad6653d0ad0373b139466fb3079d9"} Sep 29 10:59:21 crc kubenswrapper[4727]: I0929 10:59:21.381160 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 10:59:22 crc kubenswrapper[4727]: I0929 10:59:22.392262 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerStarted","Data":"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08"} Sep 29 10:59:23 crc kubenswrapper[4727]: I0929 10:59:23.405967 4727 generic.go:334] "Generic (PLEG): container finished" podID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerID="d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08" exitCode=0 Sep 29 10:59:23 crc kubenswrapper[4727]: I0929 10:59:23.406014 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerDied","Data":"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08"} Sep 29 10:59:24 crc kubenswrapper[4727]: I0929 10:59:24.416450 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerStarted","Data":"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62"} Sep 29 10:59:24 crc kubenswrapper[4727]: I0929 10:59:24.441782 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8nq2l" podStartSLOduration=2.847781656 podStartE2EDuration="5.44176036s" podCreationTimestamp="2025-09-29 10:59:19 +0000 UTC" firstStartedPulling="2025-09-29 10:59:21.380887841 +0000 UTC m=+2231.554201193" lastFinishedPulling="2025-09-29 10:59:23.974866535 +0000 UTC m=+2234.148179897" observedRunningTime="2025-09-29 10:59:24.435654484 +0000 UTC m=+2234.608967846" watchObservedRunningTime="2025-09-29 10:59:24.44176036 +0000 UTC m=+2234.615073722" Sep 29 10:59:29 crc kubenswrapper[4727]: I0929 10:59:29.842445 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:29 crc kubenswrapper[4727]: I0929 10:59:29.844016 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:29 crc kubenswrapper[4727]: I0929 10:59:29.913368 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:30 crc kubenswrapper[4727]: I0929 10:59:30.517479 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:30 crc kubenswrapper[4727]: I0929 10:59:30.567314 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:32 crc kubenswrapper[4727]: I0929 10:59:32.108357 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 10:59:32 crc kubenswrapper[4727]: E0929 10:59:32.108896 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 10:59:32 crc kubenswrapper[4727]: I0929 10:59:32.488695 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8nq2l" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="registry-server" containerID="cri-o://a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62" gracePeriod=2 Sep 29 10:59:32 crc kubenswrapper[4727]: I0929 10:59:32.936871 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.102551 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities\") pod \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.102629 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content\") pod \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.102909 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmgqg\" (UniqueName: \"kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg\") pod \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\" (UID: \"d4e8f30c-384b-42f3-b69f-a838ff2068bc\") " Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.105009 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities" (OuterVolumeSpecName: "utilities") pod "d4e8f30c-384b-42f3-b69f-a838ff2068bc" (UID: "d4e8f30c-384b-42f3-b69f-a838ff2068bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.121729 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg" (OuterVolumeSpecName: "kube-api-access-vmgqg") pod "d4e8f30c-384b-42f3-b69f-a838ff2068bc" (UID: "d4e8f30c-384b-42f3-b69f-a838ff2068bc"). InnerVolumeSpecName "kube-api-access-vmgqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.205922 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmgqg\" (UniqueName: \"kubernetes.io/projected/d4e8f30c-384b-42f3-b69f-a838ff2068bc-kube-api-access-vmgqg\") on node \"crc\" DevicePath \"\"" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.205974 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.499820 4727 generic.go:334] "Generic (PLEG): container finished" podID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerID="a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62" exitCode=0 Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.499866 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerDied","Data":"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62"} Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.499903 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nq2l" event={"ID":"d4e8f30c-384b-42f3-b69f-a838ff2068bc","Type":"ContainerDied","Data":"b50336cbe1f83c469289526b108902a5542ad6653d0ad0373b139466fb3079d9"} Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.499921 4727 scope.go:117] "RemoveContainer" containerID="a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.499925 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nq2l" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.522915 4727 scope.go:117] "RemoveContainer" containerID="d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.548783 4727 scope.go:117] "RemoveContainer" containerID="6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.595102 4727 scope.go:117] "RemoveContainer" containerID="a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62" Sep 29 10:59:33 crc kubenswrapper[4727]: E0929 10:59:33.595864 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62\": container with ID starting with a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62 not found: ID does not exist" containerID="a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.595901 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62"} err="failed to get container status \"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62\": rpc error: code = NotFound desc = could not find container \"a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62\": container with ID starting with a60b25aa00dab6f4aa8e3b345cb0feb652bb79f8eea76db24a75018d2e996a62 not found: ID does not exist" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.595924 4727 scope.go:117] "RemoveContainer" containerID="d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08" Sep 29 10:59:33 crc kubenswrapper[4727]: E0929 10:59:33.596204 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08\": container with ID starting with d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08 not found: ID does not exist" containerID="d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.596252 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08"} err="failed to get container status \"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08\": rpc error: code = NotFound desc = could not find container \"d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08\": container with ID starting with d8ba6a7cb4e4213a05c767a7cf42de4cccc1743897c9255ba7720557d5e46a08 not found: ID does not exist" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.596287 4727 scope.go:117] "RemoveContainer" containerID="6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af" Sep 29 10:59:33 crc kubenswrapper[4727]: E0929 10:59:33.596544 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af\": container with ID starting with 6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af not found: ID does not exist" containerID="6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.596570 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af"} err="failed to get container status \"6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af\": rpc error: code = NotFound desc = could not find container \"6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af\": container with ID starting with 6005e4b08fad0b52b26a4650f032355629393b4faa9c57e16204a2dcdd71e4af not found: ID does not exist" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.779372 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4e8f30c-384b-42f3-b69f-a838ff2068bc" (UID: "d4e8f30c-384b-42f3-b69f-a838ff2068bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.819236 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e8f30c-384b-42f3-b69f-a838ff2068bc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.840616 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:33 crc kubenswrapper[4727]: I0929 10:59:33.849195 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8nq2l"] Sep 29 10:59:35 crc kubenswrapper[4727]: I0929 10:59:35.123067 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" path="/var/lib/kubelet/pods/d4e8f30c-384b-42f3-b69f-a838ff2068bc/volumes" Sep 29 10:59:47 crc kubenswrapper[4727]: I0929 10:59:47.109745 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 10:59:47 crc kubenswrapper[4727]: E0929 10:59:47.111403 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.142276 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78"] Sep 29 11:00:00 crc kubenswrapper[4727]: E0929 11:00:00.143246 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="extract-utilities" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.143262 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="extract-utilities" Sep 29 11:00:00 crc kubenswrapper[4727]: E0929 11:00:00.143272 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="registry-server" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.143279 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="registry-server" Sep 29 11:00:00 crc kubenswrapper[4727]: E0929 11:00:00.143313 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="extract-content" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.143319 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="extract-content" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.143526 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4e8f30c-384b-42f3-b69f-a838ff2068bc" containerName="registry-server" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.144236 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.150016 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.151104 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.153217 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78"] Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.246754 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khx8h\" (UniqueName: \"kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.246911 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.246948 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.348135 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.348195 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.348256 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khx8h\" (UniqueName: \"kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.349439 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.357759 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.366351 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khx8h\" (UniqueName: \"kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h\") pod \"collect-profiles-29319060-q9k78\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.480279 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:00 crc kubenswrapper[4727]: I0929 11:00:00.920591 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78"] Sep 29 11:00:01 crc kubenswrapper[4727]: I0929 11:00:01.770285 4727 generic.go:334] "Generic (PLEG): container finished" podID="c27b62e6-ca28-4468-9a24-afdf664abb6c" containerID="6f5cce438e1be93a0ce42096aa85fe4e30ac3bcad3c936d842ef8c9acfa2780c" exitCode=0 Sep 29 11:00:01 crc kubenswrapper[4727]: I0929 11:00:01.770446 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" event={"ID":"c27b62e6-ca28-4468-9a24-afdf664abb6c","Type":"ContainerDied","Data":"6f5cce438e1be93a0ce42096aa85fe4e30ac3bcad3c936d842ef8c9acfa2780c"} Sep 29 11:00:01 crc kubenswrapper[4727]: I0929 11:00:01.770644 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" event={"ID":"c27b62e6-ca28-4468-9a24-afdf664abb6c","Type":"ContainerStarted","Data":"ce1087d82c7d3f196858672168a71f5732d02f91c3010d061864fa64abdb900f"} Sep 29 11:00:02 crc kubenswrapper[4727]: I0929 11:00:02.108897 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:00:02 crc kubenswrapper[4727]: E0929 11:00:02.109280 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.095418 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.204386 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume\") pod \"c27b62e6-ca28-4468-9a24-afdf664abb6c\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.204540 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume\") pod \"c27b62e6-ca28-4468-9a24-afdf664abb6c\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.204605 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khx8h\" (UniqueName: \"kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h\") pod \"c27b62e6-ca28-4468-9a24-afdf664abb6c\" (UID: \"c27b62e6-ca28-4468-9a24-afdf664abb6c\") " Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.205240 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume" (OuterVolumeSpecName: "config-volume") pod "c27b62e6-ca28-4468-9a24-afdf664abb6c" (UID: "c27b62e6-ca28-4468-9a24-afdf664abb6c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.211212 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h" (OuterVolumeSpecName: "kube-api-access-khx8h") pod "c27b62e6-ca28-4468-9a24-afdf664abb6c" (UID: "c27b62e6-ca28-4468-9a24-afdf664abb6c"). InnerVolumeSpecName "kube-api-access-khx8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.211327 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c27b62e6-ca28-4468-9a24-afdf664abb6c" (UID: "c27b62e6-ca28-4468-9a24-afdf664abb6c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.306620 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c27b62e6-ca28-4468-9a24-afdf664abb6c-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.306659 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c27b62e6-ca28-4468-9a24-afdf664abb6c-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.306675 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khx8h\" (UniqueName: \"kubernetes.io/projected/c27b62e6-ca28-4468-9a24-afdf664abb6c-kube-api-access-khx8h\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.788408 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" event={"ID":"c27b62e6-ca28-4468-9a24-afdf664abb6c","Type":"ContainerDied","Data":"ce1087d82c7d3f196858672168a71f5732d02f91c3010d061864fa64abdb900f"} Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.788458 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce1087d82c7d3f196858672168a71f5732d02f91c3010d061864fa64abdb900f" Sep 29 11:00:03 crc kubenswrapper[4727]: I0929 11:00:03.788461 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319060-q9k78" Sep 29 11:00:04 crc kubenswrapper[4727]: I0929 11:00:04.174065 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8"] Sep 29 11:00:04 crc kubenswrapper[4727]: I0929 11:00:04.182748 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319015-l24x8"] Sep 29 11:00:05 crc kubenswrapper[4727]: I0929 11:00:05.118456 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034c49df-ae6d-4483-b7cf-b5cd62bbb5a2" path="/var/lib/kubelet/pods/034c49df-ae6d-4483-b7cf-b5cd62bbb5a2/volumes" Sep 29 11:00:16 crc kubenswrapper[4727]: I0929 11:00:16.111275 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:00:16 crc kubenswrapper[4727]: E0929 11:00:16.113162 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:16 crc kubenswrapper[4727]: I0929 11:00:16.605480 4727 scope.go:117] "RemoveContainer" containerID="ec3c7e62a378703afcf313eb626eaecf2011a5dcc0c5e2cbe9fc52b0bd80ecdc" Sep 29 11:00:29 crc kubenswrapper[4727]: I0929 11:00:29.109069 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:00:29 crc kubenswrapper[4727]: E0929 11:00:29.109977 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:43 crc kubenswrapper[4727]: I0929 11:00:43.109764 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:00:43 crc kubenswrapper[4727]: E0929 11:00:43.111656 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:56 crc kubenswrapper[4727]: I0929 11:00:56.263720 4727 generic.go:334] "Generic (PLEG): container finished" podID="5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" containerID="96919151de6ae9292563c33960fb55b9580e09a07edb031a7b8fefcaaf7c68ee" exitCode=0 Sep 29 11:00:56 crc kubenswrapper[4727]: I0929 11:00:56.263816 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" event={"ID":"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea","Type":"ContainerDied","Data":"96919151de6ae9292563c33960fb55b9580e09a07edb031a7b8fefcaaf7c68ee"} Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.667551 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.784553 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0\") pod \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.785045 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory\") pod \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.785098 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgsmh\" (UniqueName: \"kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh\") pod \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.785161 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key\") pod \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.785291 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle\") pod \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\" (UID: \"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea\") " Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.790303 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" (UID: "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.790368 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh" (OuterVolumeSpecName: "kube-api-access-jgsmh") pod "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" (UID: "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea"). InnerVolumeSpecName "kube-api-access-jgsmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.812873 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" (UID: "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.813301 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" (UID: "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.820718 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory" (OuterVolumeSpecName: "inventory") pod "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" (UID: "5fd8b89e-ffdd-4382-9bc1-4f35d7547cea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.887511 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.887543 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgsmh\" (UniqueName: \"kubernetes.io/projected/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-kube-api-access-jgsmh\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.887553 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.887561 4727 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:57 crc kubenswrapper[4727]: I0929 11:00:57.887573 4727 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5fd8b89e-ffdd-4382-9bc1-4f35d7547cea-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.108208 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:00:58 crc kubenswrapper[4727]: E0929 11:00:58.108466 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.283653 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" event={"ID":"5fd8b89e-ffdd-4382-9bc1-4f35d7547cea","Type":"ContainerDied","Data":"92afad48873d575d1e79bf370cfd0b3ceb2917bd4755269d6361c3f43141d4b9"} Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.283899 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92afad48873d575d1e79bf370cfd0b3ceb2917bd4755269d6361c3f43141d4b9" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.283729 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xskrj" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.383578 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw"] Sep 29 11:00:58 crc kubenswrapper[4727]: E0929 11:00:58.384036 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27b62e6-ca28-4468-9a24-afdf664abb6c" containerName="collect-profiles" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.384053 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27b62e6-ca28-4468-9a24-afdf664abb6c" containerName="collect-profiles" Sep 29 11:00:58 crc kubenswrapper[4727]: E0929 11:00:58.384085 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.384094 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.384302 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27b62e6-ca28-4468-9a24-afdf664abb6c" containerName="collect-profiles" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.384319 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd8b89e-ffdd-4382-9bc1-4f35d7547cea" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.385084 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.387523 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.387581 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.387736 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.387844 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.388037 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.389900 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.389900 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.400695 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw"] Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401268 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401364 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401427 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401479 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401519 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbrl\" (UniqueName: \"kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401550 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401584 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401607 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.401850 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.502825 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.503077 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.503180 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.503312 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.503509 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.504131 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.504148 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbrl\" (UniqueName: \"kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.504290 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.504454 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.504539 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.507351 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.507457 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.509256 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.510007 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.510916 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.514394 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.517863 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.526714 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbrl\" (UniqueName: \"kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-b78rw\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:58 crc kubenswrapper[4727]: I0929 11:00:58.709135 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:00:59 crc kubenswrapper[4727]: I0929 11:00:59.344750 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw"] Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.135451 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29319061-6r9mf"] Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.137325 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.185525 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319061-6r9mf"] Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.312229 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" event={"ID":"a3618996-7c1c-480d-b868-25e17445d8a0","Type":"ContainerStarted","Data":"a5239518b62f1123ebcd2daf23e8c9b9d9445757e525434c5aab1b73f6746d9f"} Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.312278 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" event={"ID":"a3618996-7c1c-480d-b868-25e17445d8a0","Type":"ContainerStarted","Data":"a1fef0f1a8f225d4149ac2649504a5d04ead0266170af5d08ec2d1d810ae8c35"} Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.337126 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" podStartSLOduration=1.921273871 podStartE2EDuration="2.3371097s" podCreationTimestamp="2025-09-29 11:00:58 +0000 UTC" firstStartedPulling="2025-09-29 11:00:59.349205408 +0000 UTC m=+2329.522518770" lastFinishedPulling="2025-09-29 11:00:59.765041237 +0000 UTC m=+2329.938354599" observedRunningTime="2025-09-29 11:01:00.328442299 +0000 UTC m=+2330.501755661" watchObservedRunningTime="2025-09-29 11:01:00.3371097 +0000 UTC m=+2330.510423062" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.338178 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.338234 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.338379 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc697\" (UniqueName: \"kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.338478 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.440370 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.440477 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc697\" (UniqueName: \"kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.440518 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.440814 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.446560 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.451250 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.452155 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.461054 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc697\" (UniqueName: \"kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697\") pod \"keystone-cron-29319061-6r9mf\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.495836 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:00 crc kubenswrapper[4727]: I0929 11:01:00.970941 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29319061-6r9mf"] Sep 29 11:01:01 crc kubenswrapper[4727]: I0929 11:01:01.324562 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319061-6r9mf" event={"ID":"de7eade9-d522-4189-8a6c-60c26f0fcec7","Type":"ContainerStarted","Data":"79eca49c38f7cd8865a2d4e510d8b722e8a10f2719698229fac59ddfee5a726f"} Sep 29 11:01:01 crc kubenswrapper[4727]: I0929 11:01:01.324971 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319061-6r9mf" event={"ID":"de7eade9-d522-4189-8a6c-60c26f0fcec7","Type":"ContainerStarted","Data":"49bef0e4c13b2162ec2612c4eb0c98517684fc63566b2ea3604b4bb833695eb8"} Sep 29 11:01:01 crc kubenswrapper[4727]: I0929 11:01:01.352616 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29319061-6r9mf" podStartSLOduration=1.352588467 podStartE2EDuration="1.352588467s" podCreationTimestamp="2025-09-29 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:01:01.341662388 +0000 UTC m=+2331.514975750" watchObservedRunningTime="2025-09-29 11:01:01.352588467 +0000 UTC m=+2331.525901869" Sep 29 11:01:03 crc kubenswrapper[4727]: I0929 11:01:03.342154 4727 generic.go:334] "Generic (PLEG): container finished" podID="de7eade9-d522-4189-8a6c-60c26f0fcec7" containerID="79eca49c38f7cd8865a2d4e510d8b722e8a10f2719698229fac59ddfee5a726f" exitCode=0 Sep 29 11:01:03 crc kubenswrapper[4727]: I0929 11:01:03.342205 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319061-6r9mf" event={"ID":"de7eade9-d522-4189-8a6c-60c26f0fcec7","Type":"ContainerDied","Data":"79eca49c38f7cd8865a2d4e510d8b722e8a10f2719698229fac59ddfee5a726f"} Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.673890 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.831581 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data\") pod \"de7eade9-d522-4189-8a6c-60c26f0fcec7\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.831622 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle\") pod \"de7eade9-d522-4189-8a6c-60c26f0fcec7\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.831790 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc697\" (UniqueName: \"kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697\") pod \"de7eade9-d522-4189-8a6c-60c26f0fcec7\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.831855 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys\") pod \"de7eade9-d522-4189-8a6c-60c26f0fcec7\" (UID: \"de7eade9-d522-4189-8a6c-60c26f0fcec7\") " Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.837043 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "de7eade9-d522-4189-8a6c-60c26f0fcec7" (UID: "de7eade9-d522-4189-8a6c-60c26f0fcec7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.837536 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697" (OuterVolumeSpecName: "kube-api-access-dc697") pod "de7eade9-d522-4189-8a6c-60c26f0fcec7" (UID: "de7eade9-d522-4189-8a6c-60c26f0fcec7"). InnerVolumeSpecName "kube-api-access-dc697". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.861000 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de7eade9-d522-4189-8a6c-60c26f0fcec7" (UID: "de7eade9-d522-4189-8a6c-60c26f0fcec7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.902416 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data" (OuterVolumeSpecName: "config-data") pod "de7eade9-d522-4189-8a6c-60c26f0fcec7" (UID: "de7eade9-d522-4189-8a6c-60c26f0fcec7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.934970 4727 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.935368 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.935380 4727 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7eade9-d522-4189-8a6c-60c26f0fcec7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 11:01:04 crc kubenswrapper[4727]: I0929 11:01:04.935394 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc697\" (UniqueName: \"kubernetes.io/projected/de7eade9-d522-4189-8a6c-60c26f0fcec7-kube-api-access-dc697\") on node \"crc\" DevicePath \"\"" Sep 29 11:01:05 crc kubenswrapper[4727]: I0929 11:01:05.365941 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29319061-6r9mf" event={"ID":"de7eade9-d522-4189-8a6c-60c26f0fcec7","Type":"ContainerDied","Data":"49bef0e4c13b2162ec2612c4eb0c98517684fc63566b2ea3604b4bb833695eb8"} Sep 29 11:01:05 crc kubenswrapper[4727]: I0929 11:01:05.365996 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49bef0e4c13b2162ec2612c4eb0c98517684fc63566b2ea3604b4bb833695eb8" Sep 29 11:01:05 crc kubenswrapper[4727]: I0929 11:01:05.365998 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29319061-6r9mf" Sep 29 11:01:11 crc kubenswrapper[4727]: I0929 11:01:11.115857 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:01:11 crc kubenswrapper[4727]: E0929 11:01:11.116691 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:01:25 crc kubenswrapper[4727]: I0929 11:01:25.108276 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:01:25 crc kubenswrapper[4727]: E0929 11:01:25.109251 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:01:38 crc kubenswrapper[4727]: I0929 11:01:38.107868 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:01:38 crc kubenswrapper[4727]: E0929 11:01:38.108558 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:01:53 crc kubenswrapper[4727]: I0929 11:01:53.109410 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:01:53 crc kubenswrapper[4727]: E0929 11:01:53.110480 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:02:08 crc kubenswrapper[4727]: I0929 11:02:08.109110 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:02:08 crc kubenswrapper[4727]: E0929 11:02:08.110037 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:02:20 crc kubenswrapper[4727]: I0929 11:02:20.109224 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:02:20 crc kubenswrapper[4727]: E0929 11:02:20.110106 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:02:33 crc kubenswrapper[4727]: I0929 11:02:33.109760 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:02:33 crc kubenswrapper[4727]: E0929 11:02:33.111146 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:02:44 crc kubenswrapper[4727]: I0929 11:02:44.109412 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:02:44 crc kubenswrapper[4727]: E0929 11:02:44.110761 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:02:57 crc kubenswrapper[4727]: I0929 11:02:57.109258 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:02:57 crc kubenswrapper[4727]: E0929 11:02:57.111190 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:03:09 crc kubenswrapper[4727]: I0929 11:03:09.108543 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:03:09 crc kubenswrapper[4727]: E0929 11:03:09.109238 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:03:24 crc kubenswrapper[4727]: I0929 11:03:24.108395 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:03:24 crc kubenswrapper[4727]: E0929 11:03:24.109239 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:03:36 crc kubenswrapper[4727]: I0929 11:03:36.109731 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:03:36 crc kubenswrapper[4727]: E0929 11:03:36.111920 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:03:48 crc kubenswrapper[4727]: I0929 11:03:48.108116 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:03:48 crc kubenswrapper[4727]: E0929 11:03:48.108948 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:04:01 crc kubenswrapper[4727]: I0929 11:04:01.116111 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:04:01 crc kubenswrapper[4727]: E0929 11:04:01.117038 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:04:09 crc kubenswrapper[4727]: I0929 11:04:09.077595 4727 generic.go:334] "Generic (PLEG): container finished" podID="a3618996-7c1c-480d-b868-25e17445d8a0" containerID="a5239518b62f1123ebcd2daf23e8c9b9d9445757e525434c5aab1b73f6746d9f" exitCode=0 Sep 29 11:04:09 crc kubenswrapper[4727]: I0929 11:04:09.078090 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" event={"ID":"a3618996-7c1c-480d-b868-25e17445d8a0","Type":"ContainerDied","Data":"a5239518b62f1123ebcd2daf23e8c9b9d9445757e525434c5aab1b73f6746d9f"} Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.068961 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:10 crc kubenswrapper[4727]: E0929 11:04:10.071225 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7eade9-d522-4189-8a6c-60c26f0fcec7" containerName="keystone-cron" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.071268 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7eade9-d522-4189-8a6c-60c26f0fcec7" containerName="keystone-cron" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.071590 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="de7eade9-d522-4189-8a6c-60c26f0fcec7" containerName="keystone-cron" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.073560 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.083623 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.260475 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.260573 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.260607 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fjbn\" (UniqueName: \"kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.362467 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fjbn\" (UniqueName: \"kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.362639 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.362729 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.363646 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.363675 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.400099 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fjbn\" (UniqueName: \"kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn\") pod \"community-operators-nml8l\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.607789 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.700002 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.769996 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770370 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770391 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770425 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770442 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770546 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770636 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxbrl\" (UniqueName: \"kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770824 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.770861 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle\") pod \"a3618996-7c1c-480d-b868-25e17445d8a0\" (UID: \"a3618996-7c1c-480d-b868-25e17445d8a0\") " Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.775648 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.792184 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl" (OuterVolumeSpecName: "kube-api-access-vxbrl") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "kube-api-access-vxbrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.799243 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.806274 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.810084 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory" (OuterVolumeSpecName: "inventory") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.815837 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.818251 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.818432 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.818854 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a3618996-7c1c-480d-b868-25e17445d8a0" (UID: "a3618996-7c1c-480d-b868-25e17445d8a0"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872399 4727 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872436 4727 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872451 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872463 4727 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872474 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872487 4727 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a3618996-7c1c-480d-b868-25e17445d8a0-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872499 4727 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872509 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxbrl\" (UniqueName: \"kubernetes.io/projected/a3618996-7c1c-480d-b868-25e17445d8a0-kube-api-access-vxbrl\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:10 crc kubenswrapper[4727]: I0929 11:04:10.872520 4727 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a3618996-7c1c-480d-b868-25e17445d8a0-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.096049 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" event={"ID":"a3618996-7c1c-480d-b868-25e17445d8a0","Type":"ContainerDied","Data":"a1fef0f1a8f225d4149ac2649504a5d04ead0266170af5d08ec2d1d810ae8c35"} Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.096084 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-b78rw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.096098 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1fef0f1a8f225d4149ac2649504a5d04ead0266170af5d08ec2d1d810ae8c35" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.183960 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.219676 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw"] Sep 29 11:04:11 crc kubenswrapper[4727]: E0929 11:04:11.220545 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3618996-7c1c-480d-b868-25e17445d8a0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.220564 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3618996-7c1c-480d-b868-25e17445d8a0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.221114 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3618996-7c1c-480d-b868-25e17445d8a0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.222300 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.232160 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-x8s9b" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.232403 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.233037 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.233076 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.233217 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.242841 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw"] Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302203 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302405 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302444 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302538 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92g8v\" (UniqueName: \"kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302608 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.302742 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.303241 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406093 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406189 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406229 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406358 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406388 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406441 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92g8v\" (UniqueName: \"kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.406484 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.412096 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.412096 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.413129 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.413770 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.414815 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.415472 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.427231 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92g8v\" (UniqueName: \"kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:11 crc kubenswrapper[4727]: I0929 11:04:11.574802 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.059772 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw"] Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.107019 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" event={"ID":"77fc39d5-f092-4f06-9a6b-5d156935de57","Type":"ContainerStarted","Data":"2cd378905f07be9fb793b8031e749e31b5146c967c698d0d2cdabaac8ea32e78"} Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.108039 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:04:12 crc kubenswrapper[4727]: E0929 11:04:12.108275 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.109467 4727 generic.go:334] "Generic (PLEG): container finished" podID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerID="ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc" exitCode=0 Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.109514 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerDied","Data":"ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc"} Sep 29 11:04:12 crc kubenswrapper[4727]: I0929 11:04:12.109544 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerStarted","Data":"0a16a2dbf6f3a1565b93a3a6620ec3d64bfd5400d8f85f909a1c498b45c535e2"} Sep 29 11:04:13 crc kubenswrapper[4727]: I0929 11:04:13.122677 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" event={"ID":"77fc39d5-f092-4f06-9a6b-5d156935de57","Type":"ContainerStarted","Data":"40d7e0ad3947323e5dec3c62f6eb9d17d1e12dd50ea7a1d11ac9f6582d62d646"} Sep 29 11:04:13 crc kubenswrapper[4727]: I0929 11:04:13.154053 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" podStartSLOduration=1.458137789 podStartE2EDuration="2.154032408s" podCreationTimestamp="2025-09-29 11:04:11 +0000 UTC" firstStartedPulling="2025-09-29 11:04:12.072280722 +0000 UTC m=+2522.245594084" lastFinishedPulling="2025-09-29 11:04:12.768175341 +0000 UTC m=+2522.941488703" observedRunningTime="2025-09-29 11:04:13.146421274 +0000 UTC m=+2523.319734636" watchObservedRunningTime="2025-09-29 11:04:13.154032408 +0000 UTC m=+2523.327345770" Sep 29 11:04:14 crc kubenswrapper[4727]: I0929 11:04:14.133504 4727 generic.go:334] "Generic (PLEG): container finished" podID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerID="2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454" exitCode=0 Sep 29 11:04:14 crc kubenswrapper[4727]: I0929 11:04:14.133573 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerDied","Data":"2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454"} Sep 29 11:04:15 crc kubenswrapper[4727]: I0929 11:04:15.142939 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerStarted","Data":"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4"} Sep 29 11:04:15 crc kubenswrapper[4727]: I0929 11:04:15.163117 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nml8l" podStartSLOduration=2.737294703 podStartE2EDuration="5.163099228s" podCreationTimestamp="2025-09-29 11:04:10 +0000 UTC" firstStartedPulling="2025-09-29 11:04:12.110875709 +0000 UTC m=+2522.284189071" lastFinishedPulling="2025-09-29 11:04:14.536680234 +0000 UTC m=+2524.709993596" observedRunningTime="2025-09-29 11:04:15.158327467 +0000 UTC m=+2525.331640829" watchObservedRunningTime="2025-09-29 11:04:15.163099228 +0000 UTC m=+2525.336412590" Sep 29 11:04:20 crc kubenswrapper[4727]: I0929 11:04:20.701093 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:20 crc kubenswrapper[4727]: I0929 11:04:20.701679 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:20 crc kubenswrapper[4727]: I0929 11:04:20.745885 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:21 crc kubenswrapper[4727]: I0929 11:04:21.247414 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:21 crc kubenswrapper[4727]: I0929 11:04:21.300773 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.226445 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nml8l" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="registry-server" containerID="cri-o://5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4" gracePeriod=2 Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.676906 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.772441 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fjbn\" (UniqueName: \"kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn\") pod \"41f7c437-fa90-4287-b009-61bb0d6365c3\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.772526 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities\") pod \"41f7c437-fa90-4287-b009-61bb0d6365c3\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.772724 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content\") pod \"41f7c437-fa90-4287-b009-61bb0d6365c3\" (UID: \"41f7c437-fa90-4287-b009-61bb0d6365c3\") " Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.773438 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities" (OuterVolumeSpecName: "utilities") pod "41f7c437-fa90-4287-b009-61bb0d6365c3" (UID: "41f7c437-fa90-4287-b009-61bb0d6365c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.779693 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn" (OuterVolumeSpecName: "kube-api-access-5fjbn") pod "41f7c437-fa90-4287-b009-61bb0d6365c3" (UID: "41f7c437-fa90-4287-b009-61bb0d6365c3"). InnerVolumeSpecName "kube-api-access-5fjbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.824029 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41f7c437-fa90-4287-b009-61bb0d6365c3" (UID: "41f7c437-fa90-4287-b009-61bb0d6365c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.875813 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.875846 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41f7c437-fa90-4287-b009-61bb0d6365c3-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:23 crc kubenswrapper[4727]: I0929 11:04:23.875860 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fjbn\" (UniqueName: \"kubernetes.io/projected/41f7c437-fa90-4287-b009-61bb0d6365c3-kube-api-access-5fjbn\") on node \"crc\" DevicePath \"\"" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.108805 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.239712 4727 generic.go:334] "Generic (PLEG): container finished" podID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerID="5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4" exitCode=0 Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.239749 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerDied","Data":"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4"} Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.239920 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nml8l" event={"ID":"41f7c437-fa90-4287-b009-61bb0d6365c3","Type":"ContainerDied","Data":"0a16a2dbf6f3a1565b93a3a6620ec3d64bfd5400d8f85f909a1c498b45c535e2"} Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.239940 4727 scope.go:117] "RemoveContainer" containerID="5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.239793 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nml8l" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.267930 4727 scope.go:117] "RemoveContainer" containerID="2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.273816 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.283390 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nml8l"] Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.303266 4727 scope.go:117] "RemoveContainer" containerID="ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.323925 4727 scope.go:117] "RemoveContainer" containerID="5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4" Sep 29 11:04:24 crc kubenswrapper[4727]: E0929 11:04:24.324370 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4\": container with ID starting with 5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4 not found: ID does not exist" containerID="5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.324423 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4"} err="failed to get container status \"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4\": rpc error: code = NotFound desc = could not find container \"5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4\": container with ID starting with 5aa56b67076d12e5bebd083daa558eea775d44d281f5552655adf82184bc00b4 not found: ID does not exist" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.324456 4727 scope.go:117] "RemoveContainer" containerID="2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454" Sep 29 11:04:24 crc kubenswrapper[4727]: E0929 11:04:24.324850 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454\": container with ID starting with 2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454 not found: ID does not exist" containerID="2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.324883 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454"} err="failed to get container status \"2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454\": rpc error: code = NotFound desc = could not find container \"2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454\": container with ID starting with 2001a38737c3bf444319b51973a6392dde3c808715c001fb812f396ad435a454 not found: ID does not exist" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.324906 4727 scope.go:117] "RemoveContainer" containerID="ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc" Sep 29 11:04:24 crc kubenswrapper[4727]: E0929 11:04:24.325195 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc\": container with ID starting with ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc not found: ID does not exist" containerID="ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc" Sep 29 11:04:24 crc kubenswrapper[4727]: I0929 11:04:24.325301 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc"} err="failed to get container status \"ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc\": rpc error: code = NotFound desc = could not find container \"ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc\": container with ID starting with ee3ead08a02650b30a9c2bee73fb3887053907ae87cdbe0222fd106c1e4559fc not found: ID does not exist" Sep 29 11:04:25 crc kubenswrapper[4727]: I0929 11:04:25.124649 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" path="/var/lib/kubelet/pods/41f7c437-fa90-4287-b009-61bb0d6365c3/volumes" Sep 29 11:04:25 crc kubenswrapper[4727]: I0929 11:04:25.253277 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55"} Sep 29 11:06:29 crc kubenswrapper[4727]: I0929 11:06:29.406387 4727 generic.go:334] "Generic (PLEG): container finished" podID="77fc39d5-f092-4f06-9a6b-5d156935de57" containerID="40d7e0ad3947323e5dec3c62f6eb9d17d1e12dd50ea7a1d11ac9f6582d62d646" exitCode=0 Sep 29 11:06:29 crc kubenswrapper[4727]: I0929 11:06:29.406460 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" event={"ID":"77fc39d5-f092-4f06-9a6b-5d156935de57","Type":"ContainerDied","Data":"40d7e0ad3947323e5dec3c62f6eb9d17d1e12dd50ea7a1d11ac9f6582d62d646"} Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.798526 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.985853 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986047 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986220 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986264 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986496 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986593 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92g8v\" (UniqueName: \"kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.986635 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle\") pod \"77fc39d5-f092-4f06-9a6b-5d156935de57\" (UID: \"77fc39d5-f092-4f06-9a6b-5d156935de57\") " Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.993615 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:30 crc kubenswrapper[4727]: I0929 11:06:30.994269 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v" (OuterVolumeSpecName: "kube-api-access-92g8v") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "kube-api-access-92g8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.018608 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.020810 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.023536 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory" (OuterVolumeSpecName: "inventory") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.024311 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.036093 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "77fc39d5-f092-4f06-9a6b-5d156935de57" (UID: "77fc39d5-f092-4f06-9a6b-5d156935de57"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.089427 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.089592 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92g8v\" (UniqueName: \"kubernetes.io/projected/77fc39d5-f092-4f06-9a6b-5d156935de57-kube-api-access-92g8v\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.089691 4727 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.089887 4727 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.090005 4727 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.090087 4727 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-inventory\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.099401 4727 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77fc39d5-f092-4f06-9a6b-5d156935de57-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.425880 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" event={"ID":"77fc39d5-f092-4f06-9a6b-5d156935de57","Type":"ContainerDied","Data":"2cd378905f07be9fb793b8031e749e31b5146c967c698d0d2cdabaac8ea32e78"} Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.426147 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cd378905f07be9fb793b8031e749e31b5146c967c698d0d2cdabaac8ea32e78" Sep 29 11:06:31 crc kubenswrapper[4727]: I0929 11:06:31.425963 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw" Sep 29 11:06:49 crc kubenswrapper[4727]: I0929 11:06:49.246433 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:06:49 crc kubenswrapper[4727]: I0929 11:06:49.247005 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.838562 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Sep 29 11:07:15 crc kubenswrapper[4727]: E0929 11:07:15.839476 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77fc39d5-f092-4f06-9a6b-5d156935de57" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839494 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="77fc39d5-f092-4f06-9a6b-5d156935de57" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 29 11:07:15 crc kubenswrapper[4727]: E0929 11:07:15.839529 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="extract-content" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839535 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="extract-content" Sep 29 11:07:15 crc kubenswrapper[4727]: E0929 11:07:15.839723 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="registry-server" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839734 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="registry-server" Sep 29 11:07:15 crc kubenswrapper[4727]: E0929 11:07:15.839750 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="extract-utilities" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839756 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="extract-utilities" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839951 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="77fc39d5-f092-4f06-9a6b-5d156935de57" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.839970 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f7c437-fa90-4287-b009-61bb0d6365c3" containerName="registry-server" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.840621 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.842864 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.843223 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-l4rvj" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.847893 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.849281 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.860519 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888188 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c4m4\" (UniqueName: \"kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888254 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888285 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888323 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888360 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888381 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888427 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888445 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.888463 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.990856 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c4m4\" (UniqueName: \"kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.990920 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.990954 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991003 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991038 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991066 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991132 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991159 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.991225 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.992041 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.992501 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.992636 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.993202 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.996356 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:15 crc kubenswrapper[4727]: I0929 11:07:15.997938 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.001715 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.002729 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.011129 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c4m4\" (UniqueName: \"kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.020955 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.164414 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.644991 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 29 11:07:16 crc kubenswrapper[4727]: W0929 11:07:16.650716 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e997369_3d49_4bc4_95ec_77c503cda426.slice/crio-c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac WatchSource:0}: Error finding container c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac: Status 404 returned error can't find the container with id c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.653436 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 11:07:16 crc kubenswrapper[4727]: I0929 11:07:16.861484 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0e997369-3d49-4bc4-95ec-77c503cda426","Type":"ContainerStarted","Data":"c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac"} Sep 29 11:07:19 crc kubenswrapper[4727]: I0929 11:07:19.246867 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:07:19 crc kubenswrapper[4727]: I0929 11:07:19.247241 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:07:42 crc kubenswrapper[4727]: E0929 11:07:42.860800 4727 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Sep 29 11:07:42 crc kubenswrapper[4727]: E0929 11:07:42.861650 4727 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5c4m4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(0e997369-3d49-4bc4-95ec-77c503cda426): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 29 11:07:42 crc kubenswrapper[4727]: E0929 11:07:42.862953 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="0e997369-3d49-4bc4-95ec-77c503cda426" Sep 29 11:07:43 crc kubenswrapper[4727]: E0929 11:07:43.091583 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="0e997369-3d49-4bc4-95ec-77c503cda426" Sep 29 11:07:49 crc kubenswrapper[4727]: I0929 11:07:49.246403 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:07:49 crc kubenswrapper[4727]: I0929 11:07:49.246950 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:07:49 crc kubenswrapper[4727]: I0929 11:07:49.246992 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 11:07:49 crc kubenswrapper[4727]: I0929 11:07:49.247713 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 11:07:49 crc kubenswrapper[4727]: I0929 11:07:49.247768 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55" gracePeriod=600 Sep 29 11:07:50 crc kubenswrapper[4727]: I0929 11:07:50.168521 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55" exitCode=0 Sep 29 11:07:50 crc kubenswrapper[4727]: I0929 11:07:50.168598 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55"} Sep 29 11:07:50 crc kubenswrapper[4727]: I0929 11:07:50.169064 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1"} Sep 29 11:07:50 crc kubenswrapper[4727]: I0929 11:07:50.169099 4727 scope.go:117] "RemoveContainer" containerID="5ec973ccbf3c0da3f687f3ae6b1ed5af6af53961f27bcdf8f8896b72aad37da3" Sep 29 11:07:59 crc kubenswrapper[4727]: I0929 11:07:59.253444 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0e997369-3d49-4bc4-95ec-77c503cda426","Type":"ContainerStarted","Data":"009578547d90205067fae9f62a51f57083f2575a224f4692fcd3f15ed7c89878"} Sep 29 11:07:59 crc kubenswrapper[4727]: I0929 11:07:59.276807 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.230039569 podStartE2EDuration="45.276785407s" podCreationTimestamp="2025-09-29 11:07:14 +0000 UTC" firstStartedPulling="2025-09-29 11:07:16.653173328 +0000 UTC m=+2706.826486690" lastFinishedPulling="2025-09-29 11:07:57.699919166 +0000 UTC m=+2747.873232528" observedRunningTime="2025-09-29 11:07:59.274480959 +0000 UTC m=+2749.447794331" watchObservedRunningTime="2025-09-29 11:07:59.276785407 +0000 UTC m=+2749.450098769" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.169626 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.172797 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.189519 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.317579 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.317689 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.318156 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n75zq\" (UniqueName: \"kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.375326 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.377202 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.400788 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.420659 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.420724 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.420801 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n75zq\" (UniqueName: \"kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.421567 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.421800 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.445600 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n75zq\" (UniqueName: \"kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq\") pod \"redhat-marketplace-5sw7n\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.493759 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.523164 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.523228 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8hr\" (UniqueName: \"kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.523319 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.624706 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.626028 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.626154 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8hr\" (UniqueName: \"kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.625796 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.627053 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.654025 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8hr\" (UniqueName: \"kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr\") pod \"redhat-operators-wsvnq\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:12 crc kubenswrapper[4727]: I0929 11:08:12.693084 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:13 crc kubenswrapper[4727]: I0929 11:08:13.256902 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:13 crc kubenswrapper[4727]: W0929 11:08:13.262000 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8725e25e_cc7d_476f_bfed_ab6a088224b2.slice/crio-b8fb05ab3e20b1e4b84ad4ade81216d2c27ca65429044a9857187211d1d5f35e WatchSource:0}: Error finding container b8fb05ab3e20b1e4b84ad4ade81216d2c27ca65429044a9857187211d1d5f35e: Status 404 returned error can't find the container with id b8fb05ab3e20b1e4b84ad4ade81216d2c27ca65429044a9857187211d1d5f35e Sep 29 11:08:13 crc kubenswrapper[4727]: I0929 11:08:13.314451 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:13 crc kubenswrapper[4727]: W0929 11:08:13.317210 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod859503c2_0cf7_42ed_9eeb_39265b2165f7.slice/crio-334300be933cede376fa9e8867d35753a71f3908469cc188e8253123bfec8dd3 WatchSource:0}: Error finding container 334300be933cede376fa9e8867d35753a71f3908469cc188e8253123bfec8dd3: Status 404 returned error can't find the container with id 334300be933cede376fa9e8867d35753a71f3908469cc188e8253123bfec8dd3 Sep 29 11:08:13 crc kubenswrapper[4727]: I0929 11:08:13.375033 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerStarted","Data":"b8fb05ab3e20b1e4b84ad4ade81216d2c27ca65429044a9857187211d1d5f35e"} Sep 29 11:08:13 crc kubenswrapper[4727]: I0929 11:08:13.381511 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerStarted","Data":"334300be933cede376fa9e8867d35753a71f3908469cc188e8253123bfec8dd3"} Sep 29 11:08:14 crc kubenswrapper[4727]: I0929 11:08:14.393000 4727 generic.go:334] "Generic (PLEG): container finished" podID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerID="bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293" exitCode=0 Sep 29 11:08:14 crc kubenswrapper[4727]: I0929 11:08:14.393070 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerDied","Data":"bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293"} Sep 29 11:08:14 crc kubenswrapper[4727]: I0929 11:08:14.395430 4727 generic.go:334] "Generic (PLEG): container finished" podID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerID="f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766" exitCode=0 Sep 29 11:08:14 crc kubenswrapper[4727]: I0929 11:08:14.395643 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerDied","Data":"f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766"} Sep 29 11:08:15 crc kubenswrapper[4727]: I0929 11:08:15.412892 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerStarted","Data":"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831"} Sep 29 11:08:15 crc kubenswrapper[4727]: I0929 11:08:15.418165 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerStarted","Data":"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e"} Sep 29 11:08:16 crc kubenswrapper[4727]: I0929 11:08:16.428785 4727 generic.go:334] "Generic (PLEG): container finished" podID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerID="2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831" exitCode=0 Sep 29 11:08:16 crc kubenswrapper[4727]: I0929 11:08:16.428875 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerDied","Data":"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831"} Sep 29 11:08:16 crc kubenswrapper[4727]: I0929 11:08:16.433565 4727 generic.go:334] "Generic (PLEG): container finished" podID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerID="97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e" exitCode=0 Sep 29 11:08:16 crc kubenswrapper[4727]: I0929 11:08:16.433600 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerDied","Data":"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e"} Sep 29 11:08:17 crc kubenswrapper[4727]: I0929 11:08:17.442613 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerStarted","Data":"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35"} Sep 29 11:08:17 crc kubenswrapper[4727]: I0929 11:08:17.445092 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerStarted","Data":"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e"} Sep 29 11:08:17 crc kubenswrapper[4727]: I0929 11:08:17.463831 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5sw7n" podStartSLOduration=2.695889037 podStartE2EDuration="5.463814044s" podCreationTimestamp="2025-09-29 11:08:12 +0000 UTC" firstStartedPulling="2025-09-29 11:08:14.396303747 +0000 UTC m=+2764.569617109" lastFinishedPulling="2025-09-29 11:08:17.164228754 +0000 UTC m=+2767.337542116" observedRunningTime="2025-09-29 11:08:17.456276853 +0000 UTC m=+2767.629590215" watchObservedRunningTime="2025-09-29 11:08:17.463814044 +0000 UTC m=+2767.637127406" Sep 29 11:08:17 crc kubenswrapper[4727]: I0929 11:08:17.475719 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wsvnq" podStartSLOduration=2.805194032 podStartE2EDuration="5.475704097s" podCreationTimestamp="2025-09-29 11:08:12 +0000 UTC" firstStartedPulling="2025-09-29 11:08:14.394782399 +0000 UTC m=+2764.568095761" lastFinishedPulling="2025-09-29 11:08:17.065292474 +0000 UTC m=+2767.238605826" observedRunningTime="2025-09-29 11:08:17.472732972 +0000 UTC m=+2767.646046334" watchObservedRunningTime="2025-09-29 11:08:17.475704097 +0000 UTC m=+2767.649017459" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.496091 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.496732 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.547776 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.598053 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.694003 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.694041 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.747444 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:22 crc kubenswrapper[4727]: I0929 11:08:22.965672 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:23 crc kubenswrapper[4727]: I0929 11:08:23.560000 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:24 crc kubenswrapper[4727]: I0929 11:08:24.526022 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5sw7n" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="registry-server" containerID="cri-o://63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35" gracePeriod=2 Sep 29 11:08:24 crc kubenswrapper[4727]: I0929 11:08:24.958357 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.075143 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n75zq\" (UniqueName: \"kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq\") pod \"8725e25e-cc7d-476f-bfed-ab6a088224b2\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.075501 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content\") pod \"8725e25e-cc7d-476f-bfed-ab6a088224b2\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.075778 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities\") pod \"8725e25e-cc7d-476f-bfed-ab6a088224b2\" (UID: \"8725e25e-cc7d-476f-bfed-ab6a088224b2\") " Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.076223 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities" (OuterVolumeSpecName: "utilities") pod "8725e25e-cc7d-476f-bfed-ab6a088224b2" (UID: "8725e25e-cc7d-476f-bfed-ab6a088224b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.076724 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.081380 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq" (OuterVolumeSpecName: "kube-api-access-n75zq") pod "8725e25e-cc7d-476f-bfed-ab6a088224b2" (UID: "8725e25e-cc7d-476f-bfed-ab6a088224b2"). InnerVolumeSpecName "kube-api-access-n75zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.087725 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8725e25e-cc7d-476f-bfed-ab6a088224b2" (UID: "8725e25e-cc7d-476f-bfed-ab6a088224b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.161162 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.178955 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n75zq\" (UniqueName: \"kubernetes.io/projected/8725e25e-cc7d-476f-bfed-ab6a088224b2-kube-api-access-n75zq\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.179002 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8725e25e-cc7d-476f-bfed-ab6a088224b2-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.547364 4727 generic.go:334] "Generic (PLEG): container finished" podID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerID="63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35" exitCode=0 Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.547460 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sw7n" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.547497 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerDied","Data":"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35"} Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.547525 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sw7n" event={"ID":"8725e25e-cc7d-476f-bfed-ab6a088224b2","Type":"ContainerDied","Data":"b8fb05ab3e20b1e4b84ad4ade81216d2c27ca65429044a9857187211d1d5f35e"} Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.547545 4727 scope.go:117] "RemoveContainer" containerID="63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.580801 4727 scope.go:117] "RemoveContainer" containerID="2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.582166 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.610877 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sw7n"] Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.707762 4727 scope.go:117] "RemoveContainer" containerID="f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.769378 4727 scope.go:117] "RemoveContainer" containerID="63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35" Sep 29 11:08:25 crc kubenswrapper[4727]: E0929 11:08:25.769935 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35\": container with ID starting with 63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35 not found: ID does not exist" containerID="63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.769964 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35"} err="failed to get container status \"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35\": rpc error: code = NotFound desc = could not find container \"63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35\": container with ID starting with 63696d413ed26444e33bd01072aa7803e42961062167a6c84789eb125a190b35 not found: ID does not exist" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.769983 4727 scope.go:117] "RemoveContainer" containerID="2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831" Sep 29 11:08:25 crc kubenswrapper[4727]: E0929 11:08:25.770974 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831\": container with ID starting with 2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831 not found: ID does not exist" containerID="2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.771023 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831"} err="failed to get container status \"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831\": rpc error: code = NotFound desc = could not find container \"2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831\": container with ID starting with 2d5d6973fde0bda19ee9091ba1e1bd33b5d480b51a4935fe10df0de90f6b0831 not found: ID does not exist" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.771038 4727 scope.go:117] "RemoveContainer" containerID="f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766" Sep 29 11:08:25 crc kubenswrapper[4727]: E0929 11:08:25.771356 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766\": container with ID starting with f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766 not found: ID does not exist" containerID="f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766" Sep 29 11:08:25 crc kubenswrapper[4727]: I0929 11:08:25.771388 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766"} err="failed to get container status \"f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766\": rpc error: code = NotFound desc = could not find container \"f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766\": container with ID starting with f0478d67fa0c329255363a3792ebdd4b1052843c96ae1a61aa78013163af1766 not found: ID does not exist" Sep 29 11:08:26 crc kubenswrapper[4727]: I0929 11:08:26.557292 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wsvnq" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="registry-server" containerID="cri-o://ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e" gracePeriod=2 Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.047261 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.118826 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content\") pod \"859503c2-0cf7-42ed-9eeb-39265b2165f7\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.119017 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht8hr\" (UniqueName: \"kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr\") pod \"859503c2-0cf7-42ed-9eeb-39265b2165f7\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.119042 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities\") pod \"859503c2-0cf7-42ed-9eeb-39265b2165f7\" (UID: \"859503c2-0cf7-42ed-9eeb-39265b2165f7\") " Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.119853 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities" (OuterVolumeSpecName: "utilities") pod "859503c2-0cf7-42ed-9eeb-39265b2165f7" (UID: "859503c2-0cf7-42ed-9eeb-39265b2165f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.124686 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr" (OuterVolumeSpecName: "kube-api-access-ht8hr") pod "859503c2-0cf7-42ed-9eeb-39265b2165f7" (UID: "859503c2-0cf7-42ed-9eeb-39265b2165f7"). InnerVolumeSpecName "kube-api-access-ht8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.125974 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" path="/var/lib/kubelet/pods/8725e25e-cc7d-476f-bfed-ab6a088224b2/volumes" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.200777 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "859503c2-0cf7-42ed-9eeb-39265b2165f7" (UID: "859503c2-0cf7-42ed-9eeb-39265b2165f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.221675 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht8hr\" (UniqueName: \"kubernetes.io/projected/859503c2-0cf7-42ed-9eeb-39265b2165f7-kube-api-access-ht8hr\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.221870 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.221932 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859503c2-0cf7-42ed-9eeb-39265b2165f7-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.567221 4727 generic.go:334] "Generic (PLEG): container finished" podID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerID="ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e" exitCode=0 Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.567318 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsvnq" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.567312 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerDied","Data":"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e"} Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.567746 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsvnq" event={"ID":"859503c2-0cf7-42ed-9eeb-39265b2165f7","Type":"ContainerDied","Data":"334300be933cede376fa9e8867d35753a71f3908469cc188e8253123bfec8dd3"} Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.567766 4727 scope.go:117] "RemoveContainer" containerID="ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.599605 4727 scope.go:117] "RemoveContainer" containerID="97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.604426 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.612930 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wsvnq"] Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.619041 4727 scope.go:117] "RemoveContainer" containerID="bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.659808 4727 scope.go:117] "RemoveContainer" containerID="ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e" Sep 29 11:08:27 crc kubenswrapper[4727]: E0929 11:08:27.660321 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e\": container with ID starting with ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e not found: ID does not exist" containerID="ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.660371 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e"} err="failed to get container status \"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e\": rpc error: code = NotFound desc = could not find container \"ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e\": container with ID starting with ff226ff9907b3a5fbc3160c907a64811f05c8ecfbdd599c4b85b93b96431a92e not found: ID does not exist" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.660398 4727 scope.go:117] "RemoveContainer" containerID="97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e" Sep 29 11:08:27 crc kubenswrapper[4727]: E0929 11:08:27.660694 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e\": container with ID starting with 97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e not found: ID does not exist" containerID="97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.660799 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e"} err="failed to get container status \"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e\": rpc error: code = NotFound desc = could not find container \"97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e\": container with ID starting with 97b15855171038070887aa00b9db3088927404a4ddbb8ed652edf0fda53a476e not found: ID does not exist" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.660940 4727 scope.go:117] "RemoveContainer" containerID="bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293" Sep 29 11:08:27 crc kubenswrapper[4727]: E0929 11:08:27.661320 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293\": container with ID starting with bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293 not found: ID does not exist" containerID="bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293" Sep 29 11:08:27 crc kubenswrapper[4727]: I0929 11:08:27.661359 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293"} err="failed to get container status \"bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293\": rpc error: code = NotFound desc = could not find container \"bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293\": container with ID starting with bb2868fc8b7cecbbaa8f1e6431d77e528829883169e11e72a5f5e934c61bd293 not found: ID does not exist" Sep 29 11:08:29 crc kubenswrapper[4727]: I0929 11:08:29.125005 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" path="/var/lib/kubelet/pods/859503c2-0cf7-42ed-9eeb-39265b2165f7/volumes" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.209471 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211524 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="extract-content" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211552 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="extract-content" Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211588 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="extract-utilities" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211603 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="extract-utilities" Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211642 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211652 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211672 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="extract-content" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211681 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="extract-content" Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211715 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="extract-utilities" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211725 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="extract-utilities" Sep 29 11:09:41 crc kubenswrapper[4727]: E0929 11:09:41.211746 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.211754 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.212183 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="859503c2-0cf7-42ed-9eeb-39265b2165f7" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.212231 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="8725e25e-cc7d-476f-bfed-ab6a088224b2" containerName="registry-server" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.219663 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.233513 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.306265 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.306356 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhfct\" (UniqueName: \"kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.306553 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.408106 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhfct\" (UniqueName: \"kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.408214 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.408308 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.408874 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.408916 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.432308 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhfct\" (UniqueName: \"kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct\") pod \"certified-operators-jzhvd\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:41 crc kubenswrapper[4727]: I0929 11:09:41.548846 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:42 crc kubenswrapper[4727]: I0929 11:09:42.110438 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:42 crc kubenswrapper[4727]: I0929 11:09:42.223850 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerStarted","Data":"cefbdec6552d9dc705bdc3f51f08e22c0082604efe108324c0f240526aa4b5b3"} Sep 29 11:09:43 crc kubenswrapper[4727]: I0929 11:09:43.233708 4727 generic.go:334] "Generic (PLEG): container finished" podID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerID="ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c" exitCode=0 Sep 29 11:09:43 crc kubenswrapper[4727]: I0929 11:09:43.233756 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerDied","Data":"ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c"} Sep 29 11:09:44 crc kubenswrapper[4727]: I0929 11:09:44.245628 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerStarted","Data":"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27"} Sep 29 11:09:46 crc kubenswrapper[4727]: I0929 11:09:46.265581 4727 generic.go:334] "Generic (PLEG): container finished" podID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerID="0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27" exitCode=0 Sep 29 11:09:46 crc kubenswrapper[4727]: I0929 11:09:46.265635 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerDied","Data":"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27"} Sep 29 11:09:47 crc kubenswrapper[4727]: I0929 11:09:47.275733 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerStarted","Data":"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305"} Sep 29 11:09:47 crc kubenswrapper[4727]: I0929 11:09:47.299098 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jzhvd" podStartSLOduration=2.826242291 podStartE2EDuration="6.299080124s" podCreationTimestamp="2025-09-29 11:09:41 +0000 UTC" firstStartedPulling="2025-09-29 11:09:43.236679975 +0000 UTC m=+2853.409993337" lastFinishedPulling="2025-09-29 11:09:46.709517808 +0000 UTC m=+2856.882831170" observedRunningTime="2025-09-29 11:09:47.293629884 +0000 UTC m=+2857.466943266" watchObservedRunningTime="2025-09-29 11:09:47.299080124 +0000 UTC m=+2857.472393476" Sep 29 11:09:49 crc kubenswrapper[4727]: I0929 11:09:49.246846 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:09:49 crc kubenswrapper[4727]: I0929 11:09:49.247227 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:09:51 crc kubenswrapper[4727]: I0929 11:09:51.549125 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:51 crc kubenswrapper[4727]: I0929 11:09:51.549711 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:51 crc kubenswrapper[4727]: I0929 11:09:51.599530 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:52 crc kubenswrapper[4727]: I0929 11:09:52.364205 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:52 crc kubenswrapper[4727]: I0929 11:09:52.411103 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.334165 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jzhvd" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="registry-server" containerID="cri-o://438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305" gracePeriod=2 Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.793214 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.892209 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities\") pod \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.892530 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhfct\" (UniqueName: \"kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct\") pod \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.892695 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content\") pod \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\" (UID: \"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf\") " Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.893300 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities" (OuterVolumeSpecName: "utilities") pod "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" (UID: "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.897839 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct" (OuterVolumeSpecName: "kube-api-access-bhfct") pod "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" (UID: "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf"). InnerVolumeSpecName "kube-api-access-bhfct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.931617 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" (UID: "9bd9f034-2bc6-47e1-8786-1d2d911d6bbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.995455 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.995491 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhfct\" (UniqueName: \"kubernetes.io/projected/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-kube-api-access-bhfct\") on node \"crc\" DevicePath \"\"" Sep 29 11:09:54 crc kubenswrapper[4727]: I0929 11:09:54.995502 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.343059 4727 generic.go:334] "Generic (PLEG): container finished" podID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerID="438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305" exitCode=0 Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.343118 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jzhvd" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.343116 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerDied","Data":"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305"} Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.343288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jzhvd" event={"ID":"9bd9f034-2bc6-47e1-8786-1d2d911d6bbf","Type":"ContainerDied","Data":"cefbdec6552d9dc705bdc3f51f08e22c0082604efe108324c0f240526aa4b5b3"} Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.343322 4727 scope.go:117] "RemoveContainer" containerID="438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.368779 4727 scope.go:117] "RemoveContainer" containerID="0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.368935 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.379466 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jzhvd"] Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.389721 4727 scope.go:117] "RemoveContainer" containerID="ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.427995 4727 scope.go:117] "RemoveContainer" containerID="438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305" Sep 29 11:09:55 crc kubenswrapper[4727]: E0929 11:09:55.428467 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305\": container with ID starting with 438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305 not found: ID does not exist" containerID="438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.428514 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305"} err="failed to get container status \"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305\": rpc error: code = NotFound desc = could not find container \"438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305\": container with ID starting with 438c4406f572b3c473c3675550919314477def1a5d0a27ec7a7e6aa343fbb305 not found: ID does not exist" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.428551 4727 scope.go:117] "RemoveContainer" containerID="0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27" Sep 29 11:09:55 crc kubenswrapper[4727]: E0929 11:09:55.429108 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27\": container with ID starting with 0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27 not found: ID does not exist" containerID="0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.429144 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27"} err="failed to get container status \"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27\": rpc error: code = NotFound desc = could not find container \"0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27\": container with ID starting with 0610ffd9303cdef2c30e9a438360f3df0dc10c568ae8687cf3151ad136d26b27 not found: ID does not exist" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.429166 4727 scope.go:117] "RemoveContainer" containerID="ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c" Sep 29 11:09:55 crc kubenswrapper[4727]: E0929 11:09:55.429465 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c\": container with ID starting with ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c not found: ID does not exist" containerID="ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c" Sep 29 11:09:55 crc kubenswrapper[4727]: I0929 11:09:55.429494 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c"} err="failed to get container status \"ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c\": rpc error: code = NotFound desc = could not find container \"ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c\": container with ID starting with ca6d8733c2bb4240a3fbed7ed58803a6d961ad6bc8871c6cac2910c4b6a5d02c not found: ID does not exist" Sep 29 11:09:57 crc kubenswrapper[4727]: I0929 11:09:57.118709 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" path="/var/lib/kubelet/pods/9bd9f034-2bc6-47e1-8786-1d2d911d6bbf/volumes" Sep 29 11:10:19 crc kubenswrapper[4727]: I0929 11:10:19.247423 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:10:19 crc kubenswrapper[4727]: I0929 11:10:19.248259 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.247045 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.247748 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.247811 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.248569 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.248618 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" gracePeriod=600 Sep 29 11:10:49 crc kubenswrapper[4727]: E0929 11:10:49.402875 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.863508 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" exitCode=0 Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.863558 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1"} Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.863597 4727 scope.go:117] "RemoveContainer" containerID="c87f5ed599a67aa1f47133e9afd6f6a91be5fd1dc061a503737116d8d4017e55" Sep 29 11:10:49 crc kubenswrapper[4727]: I0929 11:10:49.864412 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:10:49 crc kubenswrapper[4727]: E0929 11:10:49.864788 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:11:02 crc kubenswrapper[4727]: I0929 11:11:02.108827 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:11:02 crc kubenswrapper[4727]: E0929 11:11:02.109591 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:11:14 crc kubenswrapper[4727]: I0929 11:11:14.108994 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:11:14 crc kubenswrapper[4727]: E0929 11:11:14.109953 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:11:26 crc kubenswrapper[4727]: I0929 11:11:26.108638 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:11:26 crc kubenswrapper[4727]: E0929 11:11:26.109510 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:11:41 crc kubenswrapper[4727]: I0929 11:11:41.114864 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:11:41 crc kubenswrapper[4727]: E0929 11:11:41.115864 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:11:54 crc kubenswrapper[4727]: I0929 11:11:54.109199 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:11:54 crc kubenswrapper[4727]: E0929 11:11:54.110096 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:12:05 crc kubenswrapper[4727]: I0929 11:12:05.108529 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:12:05 crc kubenswrapper[4727]: E0929 11:12:05.109354 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:12:16 crc kubenswrapper[4727]: I0929 11:12:16.109208 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:12:16 crc kubenswrapper[4727]: E0929 11:12:16.110757 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:12:29 crc kubenswrapper[4727]: I0929 11:12:29.121401 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:12:29 crc kubenswrapper[4727]: E0929 11:12:29.122968 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:12:42 crc kubenswrapper[4727]: I0929 11:12:42.109381 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:12:42 crc kubenswrapper[4727]: E0929 11:12:42.110861 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:12:53 crc kubenswrapper[4727]: I0929 11:12:53.108904 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:12:53 crc kubenswrapper[4727]: E0929 11:12:53.109680 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:13:06 crc kubenswrapper[4727]: I0929 11:13:06.109700 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:13:06 crc kubenswrapper[4727]: E0929 11:13:06.110683 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:13:19 crc kubenswrapper[4727]: I0929 11:13:19.108771 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:13:19 crc kubenswrapper[4727]: E0929 11:13:19.109946 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:13:32 crc kubenswrapper[4727]: I0929 11:13:32.108493 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:13:32 crc kubenswrapper[4727]: E0929 11:13:32.109813 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:13:45 crc kubenswrapper[4727]: I0929 11:13:45.108772 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:13:45 crc kubenswrapper[4727]: E0929 11:13:45.109782 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:13:57 crc kubenswrapper[4727]: I0929 11:13:57.108380 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:13:57 crc kubenswrapper[4727]: E0929 11:13:57.109498 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:14:08 crc kubenswrapper[4727]: I0929 11:14:08.108623 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:14:08 crc kubenswrapper[4727]: E0929 11:14:08.109894 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:14:21 crc kubenswrapper[4727]: I0929 11:14:21.116511 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:14:21 crc kubenswrapper[4727]: E0929 11:14:21.118755 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:14:33 crc kubenswrapper[4727]: I0929 11:14:33.109399 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:14:33 crc kubenswrapper[4727]: E0929 11:14:33.110224 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:14:45 crc kubenswrapper[4727]: I0929 11:14:45.108545 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:14:45 crc kubenswrapper[4727]: E0929 11:14:45.109209 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.108575 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:15:00 crc kubenswrapper[4727]: E0929 11:15:00.109508 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.189628 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5"] Sep 29 11:15:00 crc kubenswrapper[4727]: E0929 11:15:00.192200 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="registry-server" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.192409 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="registry-server" Sep 29 11:15:00 crc kubenswrapper[4727]: E0929 11:15:00.192511 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="extract-content" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.192574 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="extract-content" Sep 29 11:15:00 crc kubenswrapper[4727]: E0929 11:15:00.192647 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="extract-utilities" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.192736 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="extract-utilities" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.193050 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bd9f034-2bc6-47e1-8786-1d2d911d6bbf" containerName="registry-server" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.194876 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.197408 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.203311 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5"] Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.206654 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.361329 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.362163 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.362236 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzq9\" (UniqueName: \"kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.463763 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.464088 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.464218 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzq9\" (UniqueName: \"kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.465090 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.469944 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.488251 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzq9\" (UniqueName: \"kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9\") pod \"collect-profiles-29319075-mmgs5\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.527725 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:00 crc kubenswrapper[4727]: I0929 11:15:00.979595 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5"] Sep 29 11:15:01 crc kubenswrapper[4727]: I0929 11:15:01.261451 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" event={"ID":"060df8c2-fd68-4d94-88dd-652a1a0f7104","Type":"ContainerStarted","Data":"1fe27d5719d61464a6855c0cde9a2c2b67a021c293dc4d7332bb06670d61ef2b"} Sep 29 11:15:01 crc kubenswrapper[4727]: I0929 11:15:01.261781 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" event={"ID":"060df8c2-fd68-4d94-88dd-652a1a0f7104","Type":"ContainerStarted","Data":"f8ae2dbf6c1dbbf24dfb88d4fc643ed055deb398d6dd5db75883fdc40857a7d5"} Sep 29 11:15:01 crc kubenswrapper[4727]: I0929 11:15:01.290815 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" podStartSLOduration=1.290796201 podStartE2EDuration="1.290796201s" podCreationTimestamp="2025-09-29 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:15:01.288168463 +0000 UTC m=+3171.461481825" watchObservedRunningTime="2025-09-29 11:15:01.290796201 +0000 UTC m=+3171.464109553" Sep 29 11:15:02 crc kubenswrapper[4727]: I0929 11:15:02.270063 4727 generic.go:334] "Generic (PLEG): container finished" podID="060df8c2-fd68-4d94-88dd-652a1a0f7104" containerID="1fe27d5719d61464a6855c0cde9a2c2b67a021c293dc4d7332bb06670d61ef2b" exitCode=0 Sep 29 11:15:02 crc kubenswrapper[4727]: I0929 11:15:02.270125 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" event={"ID":"060df8c2-fd68-4d94-88dd-652a1a0f7104","Type":"ContainerDied","Data":"1fe27d5719d61464a6855c0cde9a2c2b67a021c293dc4d7332bb06670d61ef2b"} Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.610283 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.733570 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzq9\" (UniqueName: \"kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9\") pod \"060df8c2-fd68-4d94-88dd-652a1a0f7104\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.733776 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume\") pod \"060df8c2-fd68-4d94-88dd-652a1a0f7104\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.733908 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume\") pod \"060df8c2-fd68-4d94-88dd-652a1a0f7104\" (UID: \"060df8c2-fd68-4d94-88dd-652a1a0f7104\") " Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.735032 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume" (OuterVolumeSpecName: "config-volume") pod "060df8c2-fd68-4d94-88dd-652a1a0f7104" (UID: "060df8c2-fd68-4d94-88dd-652a1a0f7104"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.735289 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060df8c2-fd68-4d94-88dd-652a1a0f7104-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.745706 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "060df8c2-fd68-4d94-88dd-652a1a0f7104" (UID: "060df8c2-fd68-4d94-88dd-652a1a0f7104"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.745837 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9" (OuterVolumeSpecName: "kube-api-access-hxzq9") pod "060df8c2-fd68-4d94-88dd-652a1a0f7104" (UID: "060df8c2-fd68-4d94-88dd-652a1a0f7104"). InnerVolumeSpecName "kube-api-access-hxzq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.836423 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzq9\" (UniqueName: \"kubernetes.io/projected/060df8c2-fd68-4d94-88dd-652a1a0f7104-kube-api-access-hxzq9\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:03 crc kubenswrapper[4727]: I0929 11:15:03.836508 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060df8c2-fd68-4d94-88dd-652a1a0f7104-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:04 crc kubenswrapper[4727]: I0929 11:15:04.292269 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" event={"ID":"060df8c2-fd68-4d94-88dd-652a1a0f7104","Type":"ContainerDied","Data":"f8ae2dbf6c1dbbf24dfb88d4fc643ed055deb398d6dd5db75883fdc40857a7d5"} Sep 29 11:15:04 crc kubenswrapper[4727]: I0929 11:15:04.292361 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ae2dbf6c1dbbf24dfb88d4fc643ed055deb398d6dd5db75883fdc40857a7d5" Sep 29 11:15:04 crc kubenswrapper[4727]: I0929 11:15:04.292415 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319075-mmgs5" Sep 29 11:15:04 crc kubenswrapper[4727]: I0929 11:15:04.684204 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx"] Sep 29 11:15:04 crc kubenswrapper[4727]: I0929 11:15:04.695214 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319030-52fpx"] Sep 29 11:15:05 crc kubenswrapper[4727]: I0929 11:15:05.120196 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac7e10c-b919-4e7b-b2b2-4801fe56643b" path="/var/lib/kubelet/pods/fac7e10c-b919-4e7b-b2b2-4801fe56643b/volumes" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.233495 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:12 crc kubenswrapper[4727]: E0929 11:15:12.234633 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060df8c2-fd68-4d94-88dd-652a1a0f7104" containerName="collect-profiles" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.234651 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="060df8c2-fd68-4d94-88dd-652a1a0f7104" containerName="collect-profiles" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.234900 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="060df8c2-fd68-4d94-88dd-652a1a0f7104" containerName="collect-profiles" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.236871 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.253505 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.253594 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh7vg\" (UniqueName: \"kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.253772 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.255791 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.356395 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.356459 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh7vg\" (UniqueName: \"kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.356577 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.356888 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.357316 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.381126 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh7vg\" (UniqueName: \"kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg\") pod \"community-operators-h7gzf\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:12 crc kubenswrapper[4727]: I0929 11:15:12.584650 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:13 crc kubenswrapper[4727]: I0929 11:15:13.172458 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:13 crc kubenswrapper[4727]: I0929 11:15:13.410022 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerStarted","Data":"5f3804ae0b36b998c72bb854d342b77d6c00f7d8d7b2888a62f5f44c4c5929fd"} Sep 29 11:15:14 crc kubenswrapper[4727]: I0929 11:15:14.440784 4727 generic.go:334] "Generic (PLEG): container finished" podID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerID="5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e" exitCode=0 Sep 29 11:15:14 crc kubenswrapper[4727]: I0929 11:15:14.440868 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerDied","Data":"5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e"} Sep 29 11:15:14 crc kubenswrapper[4727]: I0929 11:15:14.443775 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 11:15:15 crc kubenswrapper[4727]: I0929 11:15:15.109939 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:15:15 crc kubenswrapper[4727]: E0929 11:15:15.110693 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:15:15 crc kubenswrapper[4727]: I0929 11:15:15.456477 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerStarted","Data":"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778"} Sep 29 11:15:16 crc kubenswrapper[4727]: I0929 11:15:16.465618 4727 generic.go:334] "Generic (PLEG): container finished" podID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerID="cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778" exitCode=0 Sep 29 11:15:16 crc kubenswrapper[4727]: I0929 11:15:16.465673 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerDied","Data":"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778"} Sep 29 11:15:17 crc kubenswrapper[4727]: I0929 11:15:17.012875 4727 scope.go:117] "RemoveContainer" containerID="bd9ec1520f3854213e8eaa06cfedbadbc7843e71065724f2db78cb0ee994a350" Sep 29 11:15:17 crc kubenswrapper[4727]: I0929 11:15:17.480455 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerStarted","Data":"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0"} Sep 29 11:15:17 crc kubenswrapper[4727]: I0929 11:15:17.509372 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h7gzf" podStartSLOduration=3.061479188 podStartE2EDuration="5.509325532s" podCreationTimestamp="2025-09-29 11:15:12 +0000 UTC" firstStartedPulling="2025-09-29 11:15:14.443514813 +0000 UTC m=+3184.616828175" lastFinishedPulling="2025-09-29 11:15:16.891361157 +0000 UTC m=+3187.064674519" observedRunningTime="2025-09-29 11:15:17.502143178 +0000 UTC m=+3187.675456570" watchObservedRunningTime="2025-09-29 11:15:17.509325532 +0000 UTC m=+3187.682638934" Sep 29 11:15:22 crc kubenswrapper[4727]: I0929 11:15:22.585369 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:22 crc kubenswrapper[4727]: I0929 11:15:22.586149 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:22 crc kubenswrapper[4727]: I0929 11:15:22.664368 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:23 crc kubenswrapper[4727]: I0929 11:15:23.627621 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:23 crc kubenswrapper[4727]: I0929 11:15:23.709425 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:25 crc kubenswrapper[4727]: I0929 11:15:25.565990 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h7gzf" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="registry-server" containerID="cri-o://090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0" gracePeriod=2 Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.118580 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.249962 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content\") pod \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.250088 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh7vg\" (UniqueName: \"kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg\") pod \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.250218 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities\") pod \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\" (UID: \"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba\") " Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.251080 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities" (OuterVolumeSpecName: "utilities") pod "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" (UID: "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.262476 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg" (OuterVolumeSpecName: "kube-api-access-qh7vg") pod "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" (UID: "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba"). InnerVolumeSpecName "kube-api-access-qh7vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.302166 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" (UID: "0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.353638 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh7vg\" (UniqueName: \"kubernetes.io/projected/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-kube-api-access-qh7vg\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.353707 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.353730 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.582640 4727 generic.go:334] "Generic (PLEG): container finished" podID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerID="090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0" exitCode=0 Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.582749 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerDied","Data":"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0"} Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.582810 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7gzf" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.582833 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7gzf" event={"ID":"0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba","Type":"ContainerDied","Data":"5f3804ae0b36b998c72bb854d342b77d6c00f7d8d7b2888a62f5f44c4c5929fd"} Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.582888 4727 scope.go:117] "RemoveContainer" containerID="090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.622569 4727 scope.go:117] "RemoveContainer" containerID="cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.644450 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.658349 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h7gzf"] Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.664950 4727 scope.go:117] "RemoveContainer" containerID="5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.696567 4727 scope.go:117] "RemoveContainer" containerID="090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0" Sep 29 11:15:26 crc kubenswrapper[4727]: E0929 11:15:26.697730 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0\": container with ID starting with 090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0 not found: ID does not exist" containerID="090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.697773 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0"} err="failed to get container status \"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0\": rpc error: code = NotFound desc = could not find container \"090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0\": container with ID starting with 090b16f6be14391f88a54953693793d714616ebf19bc4b07340c6900525947f0 not found: ID does not exist" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.697830 4727 scope.go:117] "RemoveContainer" containerID="cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778" Sep 29 11:15:26 crc kubenswrapper[4727]: E0929 11:15:26.698706 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778\": container with ID starting with cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778 not found: ID does not exist" containerID="cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.698765 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778"} err="failed to get container status \"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778\": rpc error: code = NotFound desc = could not find container \"cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778\": container with ID starting with cd1e90534a1b70f1eb758b88492540d14a4785130227140709d9ae90f3fae778 not found: ID does not exist" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.698804 4727 scope.go:117] "RemoveContainer" containerID="5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e" Sep 29 11:15:26 crc kubenswrapper[4727]: E0929 11:15:26.699353 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e\": container with ID starting with 5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e not found: ID does not exist" containerID="5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e" Sep 29 11:15:26 crc kubenswrapper[4727]: I0929 11:15:26.699386 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e"} err="failed to get container status \"5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e\": rpc error: code = NotFound desc = could not find container \"5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e\": container with ID starting with 5061d41840236f3f56abb545d34f08d26880737aaa488b975a747785fbb3808e not found: ID does not exist" Sep 29 11:15:27 crc kubenswrapper[4727]: I0929 11:15:27.134450 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" path="/var/lib/kubelet/pods/0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba/volumes" Sep 29 11:15:30 crc kubenswrapper[4727]: I0929 11:15:30.108797 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:15:30 crc kubenswrapper[4727]: E0929 11:15:30.109541 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:15:45 crc kubenswrapper[4727]: I0929 11:15:45.109442 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:15:45 crc kubenswrapper[4727]: E0929 11:15:45.110566 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:15:59 crc kubenswrapper[4727]: I0929 11:15:59.108583 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:15:59 crc kubenswrapper[4727]: I0929 11:15:59.949988 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f"} Sep 29 11:18:19 crc kubenswrapper[4727]: I0929 11:18:19.246563 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:18:19 crc kubenswrapper[4727]: I0929 11:18:19.247360 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.252229 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.254277 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.266100 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:18:49 crc kubenswrapper[4727]: E0929 11:18:49.266947 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="extract-content" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.266970 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="extract-content" Sep 29 11:18:49 crc kubenswrapper[4727]: E0929 11:18:49.266994 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="extract-utilities" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.267003 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="extract-utilities" Sep 29 11:18:49 crc kubenswrapper[4727]: E0929 11:18:49.267034 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="registry-server" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.267043 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="registry-server" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.269526 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0513ecdd-4e55-4eeb-b28d-c6fc2f2eceba" containerName="registry-server" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.272210 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.290123 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.461822 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.461946 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.461978 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8mn8\" (UniqueName: \"kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.563559 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.563667 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.563711 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mn8\" (UniqueName: \"kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.564271 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.564517 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.584823 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8mn8\" (UniqueName: \"kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8\") pod \"redhat-operators-7c58k\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:49 crc kubenswrapper[4727]: I0929 11:18:49.652609 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:50 crc kubenswrapper[4727]: I0929 11:18:50.100615 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:18:50 crc kubenswrapper[4727]: I0929 11:18:50.713467 4727 generic.go:334] "Generic (PLEG): container finished" podID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerID="c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859" exitCode=0 Sep 29 11:18:50 crc kubenswrapper[4727]: I0929 11:18:50.713561 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerDied","Data":"c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859"} Sep 29 11:18:50 crc kubenswrapper[4727]: I0929 11:18:50.713790 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerStarted","Data":"e3884c9c4e8aa5f9f224ebe75a74f8216a21df6e5c80dad3a96a6fe95de8bcf8"} Sep 29 11:18:51 crc kubenswrapper[4727]: I0929 11:18:51.726476 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerStarted","Data":"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2"} Sep 29 11:18:52 crc kubenswrapper[4727]: I0929 11:18:52.739997 4727 generic.go:334] "Generic (PLEG): container finished" podID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerID="1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2" exitCode=0 Sep 29 11:18:52 crc kubenswrapper[4727]: I0929 11:18:52.740061 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerDied","Data":"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2"} Sep 29 11:18:54 crc kubenswrapper[4727]: I0929 11:18:54.761007 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerStarted","Data":"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5"} Sep 29 11:18:54 crc kubenswrapper[4727]: I0929 11:18:54.784379 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7c58k" podStartSLOduration=3.332507552 podStartE2EDuration="5.784356775s" podCreationTimestamp="2025-09-29 11:18:49 +0000 UTC" firstStartedPulling="2025-09-29 11:18:50.716822519 +0000 UTC m=+3400.890135881" lastFinishedPulling="2025-09-29 11:18:53.168671742 +0000 UTC m=+3403.341985104" observedRunningTime="2025-09-29 11:18:54.778166326 +0000 UTC m=+3404.951479708" watchObservedRunningTime="2025-09-29 11:18:54.784356775 +0000 UTC m=+3404.957670137" Sep 29 11:18:59 crc kubenswrapper[4727]: I0929 11:18:59.652885 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:59 crc kubenswrapper[4727]: I0929 11:18:59.653641 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:59 crc kubenswrapper[4727]: I0929 11:18:59.700791 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:59 crc kubenswrapper[4727]: I0929 11:18:59.846183 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:18:59 crc kubenswrapper[4727]: I0929 11:18:59.931369 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:19:01 crc kubenswrapper[4727]: I0929 11:19:01.817635 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7c58k" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="registry-server" containerID="cri-o://fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5" gracePeriod=2 Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.259718 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.401602 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content\") pod \"38fc4dde-ab1e-4c72-9af2-983e525a3792\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.402300 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities\") pod \"38fc4dde-ab1e-4c72-9af2-983e525a3792\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.402557 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8mn8\" (UniqueName: \"kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8\") pod \"38fc4dde-ab1e-4c72-9af2-983e525a3792\" (UID: \"38fc4dde-ab1e-4c72-9af2-983e525a3792\") " Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.403067 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities" (OuterVolumeSpecName: "utilities") pod "38fc4dde-ab1e-4c72-9af2-983e525a3792" (UID: "38fc4dde-ab1e-4c72-9af2-983e525a3792"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.404617 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.408615 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8" (OuterVolumeSpecName: "kube-api-access-k8mn8") pod "38fc4dde-ab1e-4c72-9af2-983e525a3792" (UID: "38fc4dde-ab1e-4c72-9af2-983e525a3792"). InnerVolumeSpecName "kube-api-access-k8mn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.479347 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38fc4dde-ab1e-4c72-9af2-983e525a3792" (UID: "38fc4dde-ab1e-4c72-9af2-983e525a3792"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.507022 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38fc4dde-ab1e-4c72-9af2-983e525a3792-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.507069 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8mn8\" (UniqueName: \"kubernetes.io/projected/38fc4dde-ab1e-4c72-9af2-983e525a3792-kube-api-access-k8mn8\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.826508 4727 generic.go:334] "Generic (PLEG): container finished" podID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerID="fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5" exitCode=0 Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.826657 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerDied","Data":"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5"} Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.826893 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c58k" event={"ID":"38fc4dde-ab1e-4c72-9af2-983e525a3792","Type":"ContainerDied","Data":"e3884c9c4e8aa5f9f224ebe75a74f8216a21df6e5c80dad3a96a6fe95de8bcf8"} Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.826912 4727 scope.go:117] "RemoveContainer" containerID="fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.826689 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c58k" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.853110 4727 scope.go:117] "RemoveContainer" containerID="1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.865906 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.875879 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7c58k"] Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.884935 4727 scope.go:117] "RemoveContainer" containerID="c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.940882 4727 scope.go:117] "RemoveContainer" containerID="fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5" Sep 29 11:19:02 crc kubenswrapper[4727]: E0929 11:19:02.941315 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5\": container with ID starting with fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5 not found: ID does not exist" containerID="fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.941454 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5"} err="failed to get container status \"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5\": rpc error: code = NotFound desc = could not find container \"fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5\": container with ID starting with fdb266af213e88ddf783fe54200fd94958c029d99ab5508d05b7e2085239b6c5 not found: ID does not exist" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.941551 4727 scope.go:117] "RemoveContainer" containerID="1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2" Sep 29 11:19:02 crc kubenswrapper[4727]: E0929 11:19:02.941970 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2\": container with ID starting with 1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2 not found: ID does not exist" containerID="1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.942080 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2"} err="failed to get container status \"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2\": rpc error: code = NotFound desc = could not find container \"1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2\": container with ID starting with 1bfb76209004badfe35db3c6134c80e0e3c987ea2c1f0b01f131dde38e7451d2 not found: ID does not exist" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.942179 4727 scope.go:117] "RemoveContainer" containerID="c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859" Sep 29 11:19:02 crc kubenswrapper[4727]: E0929 11:19:02.942608 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859\": container with ID starting with c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859 not found: ID does not exist" containerID="c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859" Sep 29 11:19:02 crc kubenswrapper[4727]: I0929 11:19:02.942676 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859"} err="failed to get container status \"c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859\": rpc error: code = NotFound desc = could not find container \"c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859\": container with ID starting with c36fb040432e81ad379f0d95981896fa61daad7ad5c19eebc55d985ec5078859 not found: ID does not exist" Sep 29 11:19:03 crc kubenswrapper[4727]: I0929 11:19:03.123072 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" path="/var/lib/kubelet/pods/38fc4dde-ab1e-4c72-9af2-983e525a3792/volumes" Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.246317 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.246884 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.246939 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.247848 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.247938 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f" gracePeriod=600 Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.987056 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f" exitCode=0 Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.987135 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f"} Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.987790 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384"} Sep 29 11:19:19 crc kubenswrapper[4727]: I0929 11:19:19.987819 4727 scope.go:117] "RemoveContainer" containerID="f3e94c8a6ca5765d3c4d47f82cc9ef845bdbfb3f1fac37a732d0fa25af1d60e1" Sep 29 11:19:38 crc kubenswrapper[4727]: I0929 11:19:38.153883 4727 generic.go:334] "Generic (PLEG): container finished" podID="0e997369-3d49-4bc4-95ec-77c503cda426" containerID="009578547d90205067fae9f62a51f57083f2575a224f4692fcd3f15ed7c89878" exitCode=0 Sep 29 11:19:38 crc kubenswrapper[4727]: I0929 11:19:38.153990 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0e997369-3d49-4bc4-95ec-77c503cda426","Type":"ContainerDied","Data":"009578547d90205067fae9f62a51f57083f2575a224f4692fcd3f15ed7c89878"} Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.541700 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.673521 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.673998 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674156 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c4m4\" (UniqueName: \"kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674225 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674268 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674331 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674372 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674397 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674415 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret\") pod \"0e997369-3d49-4bc4-95ec-77c503cda426\" (UID: \"0e997369-3d49-4bc4-95ec-77c503cda426\") " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674577 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data" (OuterVolumeSpecName: "config-data") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.674997 4727 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-config-data\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.675030 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.680901 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.681155 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4" (OuterVolumeSpecName: "kube-api-access-5c4m4") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "kube-api-access-5c4m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.681479 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.703737 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.705063 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.705464 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.727213 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0e997369-3d49-4bc4-95ec-77c503cda426" (UID: "0e997369-3d49-4bc4-95ec-77c503cda426"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776698 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776740 4727 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776777 4727 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776787 4727 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776799 4727 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776810 4727 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e997369-3d49-4bc4-95ec-77c503cda426-ca-certs\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776818 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c4m4\" (UniqueName: \"kubernetes.io/projected/0e997369-3d49-4bc4-95ec-77c503cda426-kube-api-access-5c4m4\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.776826 4727 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e997369-3d49-4bc4-95ec-77c503cda426-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.798473 4727 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 29 11:19:39 crc kubenswrapper[4727]: I0929 11:19:39.878994 4727 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 29 11:19:40 crc kubenswrapper[4727]: I0929 11:19:40.173265 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0e997369-3d49-4bc4-95ec-77c503cda426","Type":"ContainerDied","Data":"c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac"} Sep 29 11:19:40 crc kubenswrapper[4727]: I0929 11:19:40.173321 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88b4510dec04e7e5c30859f593eb3da070ff189258927182618b159c2e0cfac" Sep 29 11:19:40 crc kubenswrapper[4727]: I0929 11:19:40.173351 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.691496 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 29 11:19:49 crc kubenswrapper[4727]: E0929 11:19:49.693655 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="extract-utilities" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.693764 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="extract-utilities" Sep 29 11:19:49 crc kubenswrapper[4727]: E0929 11:19:49.693842 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="registry-server" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.693941 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="registry-server" Sep 29 11:19:49 crc kubenswrapper[4727]: E0929 11:19:49.694015 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="extract-content" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.694079 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="extract-content" Sep 29 11:19:49 crc kubenswrapper[4727]: E0929 11:19:49.694161 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e997369-3d49-4bc4-95ec-77c503cda426" containerName="tempest-tests-tempest-tests-runner" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.694226 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e997369-3d49-4bc4-95ec-77c503cda426" containerName="tempest-tests-tempest-tests-runner" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.694511 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="38fc4dde-ab1e-4c72-9af2-983e525a3792" containerName="registry-server" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.694590 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e997369-3d49-4bc4-95ec-77c503cda426" containerName="tempest-tests-tempest-tests-runner" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.695319 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.697258 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-l4rvj" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.700435 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.768494 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.768957 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th97d\" (UniqueName: \"kubernetes.io/projected/0518c511-9f69-42dd-88e3-661f13f8f019-kube-api-access-th97d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.870630 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th97d\" (UniqueName: \"kubernetes.io/projected/0518c511-9f69-42dd-88e3-661f13f8f019-kube-api-access-th97d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.870777 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.871273 4727 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.889636 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th97d\" (UniqueName: \"kubernetes.io/projected/0518c511-9f69-42dd-88e3-661f13f8f019-kube-api-access-th97d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:49 crc kubenswrapper[4727]: I0929 11:19:49.902083 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0518c511-9f69-42dd-88e3-661f13f8f019\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.020330 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.450409 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.483069 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.485143 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.500900 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.583811 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd6rd\" (UniqueName: \"kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.584497 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.584728 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.686434 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.686636 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.686709 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd6rd\" (UniqueName: \"kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.686990 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.687077 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.705989 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd6rd\" (UniqueName: \"kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd\") pod \"certified-operators-78h6w\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:50 crc kubenswrapper[4727]: I0929 11:19:50.853544 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:19:51 crc kubenswrapper[4727]: I0929 11:19:51.271109 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0518c511-9f69-42dd-88e3-661f13f8f019","Type":"ContainerStarted","Data":"ae90f891a40436a4ee9a439a09ba533f47b50661e8c40fea341c39091fb44bd9"} Sep 29 11:19:51 crc kubenswrapper[4727]: I0929 11:19:51.366624 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:19:51 crc kubenswrapper[4727]: W0929 11:19:51.372509 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52642942_fa60_4f11_95b6_7584f219591a.slice/crio-a1131e5fb7bb89ae921e31eabc7103443abf02e9dfd882745852fca89e02774b WatchSource:0}: Error finding container a1131e5fb7bb89ae921e31eabc7103443abf02e9dfd882745852fca89e02774b: Status 404 returned error can't find the container with id a1131e5fb7bb89ae921e31eabc7103443abf02e9dfd882745852fca89e02774b Sep 29 11:19:52 crc kubenswrapper[4727]: I0929 11:19:52.282233 4727 generic.go:334] "Generic (PLEG): container finished" podID="52642942-fa60-4f11-95b6-7584f219591a" containerID="68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3" exitCode=0 Sep 29 11:19:52 crc kubenswrapper[4727]: I0929 11:19:52.282275 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerDied","Data":"68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3"} Sep 29 11:19:52 crc kubenswrapper[4727]: I0929 11:19:52.282586 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerStarted","Data":"a1131e5fb7bb89ae921e31eabc7103443abf02e9dfd882745852fca89e02774b"} Sep 29 11:19:52 crc kubenswrapper[4727]: I0929 11:19:52.290191 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0518c511-9f69-42dd-88e3-661f13f8f019","Type":"ContainerStarted","Data":"770284b935fd40f4957e64f2a717e17a5de48c7b2cfb5edef1003ecb187df3a7"} Sep 29 11:19:52 crc kubenswrapper[4727]: I0929 11:19:52.321245 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.472704999 podStartE2EDuration="3.321227437s" podCreationTimestamp="2025-09-29 11:19:49 +0000 UTC" firstStartedPulling="2025-09-29 11:19:50.447153717 +0000 UTC m=+3460.620467079" lastFinishedPulling="2025-09-29 11:19:51.295676155 +0000 UTC m=+3461.468989517" observedRunningTime="2025-09-29 11:19:52.317898302 +0000 UTC m=+3462.491211664" watchObservedRunningTime="2025-09-29 11:19:52.321227437 +0000 UTC m=+3462.494540799" Sep 29 11:19:54 crc kubenswrapper[4727]: I0929 11:19:54.310406 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerStarted","Data":"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500"} Sep 29 11:19:55 crc kubenswrapper[4727]: I0929 11:19:55.320237 4727 generic.go:334] "Generic (PLEG): container finished" podID="52642942-fa60-4f11-95b6-7584f219591a" containerID="e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500" exitCode=0 Sep 29 11:19:55 crc kubenswrapper[4727]: I0929 11:19:55.320294 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerDied","Data":"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500"} Sep 29 11:19:56 crc kubenswrapper[4727]: I0929 11:19:56.337264 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerStarted","Data":"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396"} Sep 29 11:20:00 crc kubenswrapper[4727]: I0929 11:20:00.854627 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:00 crc kubenswrapper[4727]: I0929 11:20:00.855617 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:00 crc kubenswrapper[4727]: I0929 11:20:00.922017 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:00 crc kubenswrapper[4727]: I0929 11:20:00.947775 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-78h6w" podStartSLOduration=7.496415578 podStartE2EDuration="10.947731042s" podCreationTimestamp="2025-09-29 11:19:50 +0000 UTC" firstStartedPulling="2025-09-29 11:19:52.290281491 +0000 UTC m=+3462.463594853" lastFinishedPulling="2025-09-29 11:19:55.741596945 +0000 UTC m=+3465.914910317" observedRunningTime="2025-09-29 11:19:56.354602735 +0000 UTC m=+3466.527916097" watchObservedRunningTime="2025-09-29 11:20:00.947731042 +0000 UTC m=+3471.121044444" Sep 29 11:20:01 crc kubenswrapper[4727]: I0929 11:20:01.447480 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:01 crc kubenswrapper[4727]: I0929 11:20:01.505067 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:20:03 crc kubenswrapper[4727]: I0929 11:20:03.394100 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-78h6w" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="registry-server" containerID="cri-o://22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396" gracePeriod=2 Sep 29 11:20:03 crc kubenswrapper[4727]: I0929 11:20:03.910693 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.052418 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities\") pod \"52642942-fa60-4f11-95b6-7584f219591a\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.052503 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd6rd\" (UniqueName: \"kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd\") pod \"52642942-fa60-4f11-95b6-7584f219591a\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.052618 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content\") pod \"52642942-fa60-4f11-95b6-7584f219591a\" (UID: \"52642942-fa60-4f11-95b6-7584f219591a\") " Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.053542 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities" (OuterVolumeSpecName: "utilities") pod "52642942-fa60-4f11-95b6-7584f219591a" (UID: "52642942-fa60-4f11-95b6-7584f219591a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.059517 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd" (OuterVolumeSpecName: "kube-api-access-vd6rd") pod "52642942-fa60-4f11-95b6-7584f219591a" (UID: "52642942-fa60-4f11-95b6-7584f219591a"). InnerVolumeSpecName "kube-api-access-vd6rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.121182 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52642942-fa60-4f11-95b6-7584f219591a" (UID: "52642942-fa60-4f11-95b6-7584f219591a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.155937 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.156007 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd6rd\" (UniqueName: \"kubernetes.io/projected/52642942-fa60-4f11-95b6-7584f219591a-kube-api-access-vd6rd\") on node \"crc\" DevicePath \"\"" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.156026 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52642942-fa60-4f11-95b6-7584f219591a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.404732 4727 generic.go:334] "Generic (PLEG): container finished" podID="52642942-fa60-4f11-95b6-7584f219591a" containerID="22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396" exitCode=0 Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.404841 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78h6w" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.404864 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerDied","Data":"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396"} Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.405481 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78h6w" event={"ID":"52642942-fa60-4f11-95b6-7584f219591a","Type":"ContainerDied","Data":"a1131e5fb7bb89ae921e31eabc7103443abf02e9dfd882745852fca89e02774b"} Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.405509 4727 scope.go:117] "RemoveContainer" containerID="22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.433793 4727 scope.go:117] "RemoveContainer" containerID="e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.439120 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.449319 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-78h6w"] Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.472120 4727 scope.go:117] "RemoveContainer" containerID="68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.503974 4727 scope.go:117] "RemoveContainer" containerID="22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396" Sep 29 11:20:04 crc kubenswrapper[4727]: E0929 11:20:04.504514 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396\": container with ID starting with 22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396 not found: ID does not exist" containerID="22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.504568 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396"} err="failed to get container status \"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396\": rpc error: code = NotFound desc = could not find container \"22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396\": container with ID starting with 22f081bf4b0c7879ace84f6b067d58c6167bf2ea9bcb2245ff2498f232564396 not found: ID does not exist" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.504605 4727 scope.go:117] "RemoveContainer" containerID="e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500" Sep 29 11:20:04 crc kubenswrapper[4727]: E0929 11:20:04.505003 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500\": container with ID starting with e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500 not found: ID does not exist" containerID="e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.505030 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500"} err="failed to get container status \"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500\": rpc error: code = NotFound desc = could not find container \"e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500\": container with ID starting with e74076cfe4ffcb63ec0dbc0fc924a47bca2903b5029fb57a51e670f41c550500 not found: ID does not exist" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.505048 4727 scope.go:117] "RemoveContainer" containerID="68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3" Sep 29 11:20:04 crc kubenswrapper[4727]: E0929 11:20:04.505302 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3\": container with ID starting with 68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3 not found: ID does not exist" containerID="68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3" Sep 29 11:20:04 crc kubenswrapper[4727]: I0929 11:20:04.505328 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3"} err="failed to get container status \"68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3\": rpc error: code = NotFound desc = could not find container \"68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3\": container with ID starting with 68e2166c504a669e6c1b7e4d5ba9589fe5ef2a59e25ccecdc0f76f183dd46df3 not found: ID does not exist" Sep 29 11:20:05 crc kubenswrapper[4727]: I0929 11:20:05.121351 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52642942-fa60-4f11-95b6-7584f219591a" path="/var/lib/kubelet/pods/52642942-fa60-4f11-95b6-7584f219591a/volumes" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.209918 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6h5r/must-gather-fh9v6"] Sep 29 11:20:09 crc kubenswrapper[4727]: E0929 11:20:09.210852 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="extract-utilities" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.210866 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="extract-utilities" Sep 29 11:20:09 crc kubenswrapper[4727]: E0929 11:20:09.210886 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="registry-server" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.210893 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="registry-server" Sep 29 11:20:09 crc kubenswrapper[4727]: E0929 11:20:09.210918 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="extract-content" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.210925 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="extract-content" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.211133 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="52642942-fa60-4f11-95b6-7584f219591a" containerName="registry-server" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.212146 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.230813 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6h5r/must-gather-fh9v6"] Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.232105 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t6h5r"/"default-dockercfg-mqhc7" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.232303 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6h5r"/"openshift-service-ca.crt" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.232504 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6h5r"/"kube-root-ca.crt" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.351649 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kql42\" (UniqueName: \"kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.352137 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.454276 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kql42\" (UniqueName: \"kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.454382 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.454982 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.490894 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kql42\" (UniqueName: \"kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42\") pod \"must-gather-fh9v6\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:09 crc kubenswrapper[4727]: I0929 11:20:09.535827 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:20:10 crc kubenswrapper[4727]: W0929 11:20:10.004194 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefeb8425_b618_4251_ad73_3be72e98289b.slice/crio-7d377283bd146638bfb5def6e23271766a06ce3d3973ec01a144995d3b6572c4 WatchSource:0}: Error finding container 7d377283bd146638bfb5def6e23271766a06ce3d3973ec01a144995d3b6572c4: Status 404 returned error can't find the container with id 7d377283bd146638bfb5def6e23271766a06ce3d3973ec01a144995d3b6572c4 Sep 29 11:20:10 crc kubenswrapper[4727]: I0929 11:20:10.004702 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6h5r/must-gather-fh9v6"] Sep 29 11:20:10 crc kubenswrapper[4727]: I0929 11:20:10.462664 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" event={"ID":"efeb8425-b618-4251-ad73-3be72e98289b","Type":"ContainerStarted","Data":"7d377283bd146638bfb5def6e23271766a06ce3d3973ec01a144995d3b6572c4"} Sep 29 11:20:15 crc kubenswrapper[4727]: I0929 11:20:15.536354 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" event={"ID":"efeb8425-b618-4251-ad73-3be72e98289b","Type":"ContainerStarted","Data":"16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff"} Sep 29 11:20:15 crc kubenswrapper[4727]: I0929 11:20:15.537043 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" event={"ID":"efeb8425-b618-4251-ad73-3be72e98289b","Type":"ContainerStarted","Data":"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b"} Sep 29 11:20:15 crc kubenswrapper[4727]: I0929 11:20:15.553599 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" podStartSLOduration=2.201593029 podStartE2EDuration="6.553582694s" podCreationTimestamp="2025-09-29 11:20:09 +0000 UTC" firstStartedPulling="2025-09-29 11:20:10.006681341 +0000 UTC m=+3480.179994723" lastFinishedPulling="2025-09-29 11:20:14.358670986 +0000 UTC m=+3484.531984388" observedRunningTime="2025-09-29 11:20:15.552141847 +0000 UTC m=+3485.725455209" watchObservedRunningTime="2025-09-29 11:20:15.553582694 +0000 UTC m=+3485.726896056" Sep 29 11:20:17 crc kubenswrapper[4727]: I0929 11:20:17.932933 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-jnzrt"] Sep 29 11:20:17 crc kubenswrapper[4727]: I0929 11:20:17.934598 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.037983 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxc2m\" (UniqueName: \"kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.038040 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.140500 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxc2m\" (UniqueName: \"kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.140572 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.140837 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.162555 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxc2m\" (UniqueName: \"kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m\") pod \"crc-debug-jnzrt\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.253814 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:20:18 crc kubenswrapper[4727]: W0929 11:20:18.290313 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod208c21a2_8735_43ea_9396_b35f22696f66.slice/crio-582fd956a8dbf149daa4e15061893dbf65210203c40a0398683882987d376d95 WatchSource:0}: Error finding container 582fd956a8dbf149daa4e15061893dbf65210203c40a0398683882987d376d95: Status 404 returned error can't find the container with id 582fd956a8dbf149daa4e15061893dbf65210203c40a0398683882987d376d95 Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.293502 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 11:20:18 crc kubenswrapper[4727]: I0929 11:20:18.568596 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" event={"ID":"208c21a2-8735-43ea-9396-b35f22696f66","Type":"ContainerStarted","Data":"582fd956a8dbf149daa4e15061893dbf65210203c40a0398683882987d376d95"} Sep 29 11:20:30 crc kubenswrapper[4727]: I0929 11:20:30.747552 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" event={"ID":"208c21a2-8735-43ea-9396-b35f22696f66","Type":"ContainerStarted","Data":"82e40f9791267d39a35763c93a41de667a721f0089bd428df3a540f2011ae3f8"} Sep 29 11:20:30 crc kubenswrapper[4727]: I0929 11:20:30.761547 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" podStartSLOduration=2.016687013 podStartE2EDuration="13.761520606s" podCreationTimestamp="2025-09-29 11:20:17 +0000 UTC" firstStartedPulling="2025-09-29 11:20:18.2931546 +0000 UTC m=+3488.466467962" lastFinishedPulling="2025-09-29 11:20:30.037988193 +0000 UTC m=+3500.211301555" observedRunningTime="2025-09-29 11:20:30.76091103 +0000 UTC m=+3500.934224392" watchObservedRunningTime="2025-09-29 11:20:30.761520606 +0000 UTC m=+3500.934833978" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.152735 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bc65c799d-pbjqp_7529e49e-8904-4b11-ba31-ed805fc0fee2/barbican-api-log/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.169802 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bc65c799d-pbjqp_7529e49e-8904-4b11-ba31-ed805fc0fee2/barbican-api/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.246673 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.246722 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.356199 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b58bf9d-2f9w7_978ca6db-0819-4482-a407-a11ea6ebc3f2/barbican-keystone-listener/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.453551 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b58bf9d-2f9w7_978ca6db-0819-4482-a407-a11ea6ebc3f2/barbican-keystone-listener-log/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.571515 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb4b7877c-kksr7_134d320a-33b6-4df1-bf0e-da396f28f90e/barbican-worker/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.669555 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb4b7877c-kksr7_134d320a-33b6-4df1-bf0e-da396f28f90e/barbican-worker-log/0.log" Sep 29 11:21:19 crc kubenswrapper[4727]: I0929 11:21:19.803943 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl_21709a4b-d29f-47a8-acf8-ba9c1e2415c3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.131247 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/ceilometer-notification-agent/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.135762 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/ceilometer-central-agent/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.305446 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/proxy-httpd/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.312931 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/sg-core/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.498266 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0248c486-ae28-4a7c-96f6-e97cde83ae7b/cinder-api-log/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.527803 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0248c486-ae28-4a7c-96f6-e97cde83ae7b/cinder-api/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.747253 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ae9bf7b1-b3a8-4b34-8930-3751c0230705/probe/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.751036 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ae9bf7b1-b3a8-4b34-8930-3751c0230705/cinder-scheduler/0.log" Sep 29 11:21:20 crc kubenswrapper[4727]: I0929 11:21:20.982699 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zr96k_e626458d-df81-47e7-aa33-47f32d67759d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.095250 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr_d1b76565-6e67-46cf-9cc6-0f1b70d814a6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.195201 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/init/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.357899 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/init/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.478406 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/dnsmasq-dns/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.641154 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7_96f1c41e-9277-4df7-afd8-711f2efabcf6/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.779585 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_56bbd245-faf2-4c21-b4c9-172a501374fd/glance-httpd/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.810913 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_56bbd245-faf2-4c21-b4c9-172a501374fd/glance-log/0.log" Sep 29 11:21:21 crc kubenswrapper[4727]: I0929 11:21:21.983976 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b1ab9d58-dea7-47fc-a259-5b9b8f7a7230/glance-log/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.029687 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b1ab9d58-dea7-47fc-a259-5b9b8f7a7230/glance-httpd/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.407805 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-54cbdfc6b8-xl7cd_4b7b8921-ae97-45b0-a082-07db49561514/horizon/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.535522 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xqw72_d3ec4e67-b454-43bd-ab50-0257dc2e517d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.698441 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-54cbdfc6b8-xl7cd_4b7b8921-ae97-45b0-a082-07db49561514/horizon-log/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.700428 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-tzhl4_99127a9e-ae57-42b7-9392-e3f3d7ab43ff/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.973089 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64f85d6496-qmbdm_9eae012b-b92e-44ae-8396-f7331f267787/keystone-api/0.log" Sep 29 11:21:22 crc kubenswrapper[4727]: I0929 11:21:22.979051 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29319061-6r9mf_de7eade9-d522-4189-8a6c-60c26f0fcec7/keystone-cron/0.log" Sep 29 11:21:23 crc kubenswrapper[4727]: I0929 11:21:23.163687 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_bbc0072e-f90a-4518-8505-bee6f68abbaa/kube-state-metrics/0.log" Sep 29 11:21:23 crc kubenswrapper[4727]: I0929 11:21:23.209205 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xskrj_5fd8b89e-ffdd-4382-9bc1-4f35d7547cea/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:23 crc kubenswrapper[4727]: I0929 11:21:23.600233 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68ff49cc97-zxm88_326c2e7e-51f1-4f8c-9d91-4572393d8c11/neutron-api/0.log" Sep 29 11:21:23 crc kubenswrapper[4727]: I0929 11:21:23.601740 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68ff49cc97-zxm88_326c2e7e-51f1-4f8c-9d91-4572393d8c11/neutron-httpd/0.log" Sep 29 11:21:23 crc kubenswrapper[4727]: I0929 11:21:23.846163 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t_4bd4cabe-bcac-4332-b7be-326693d4bc95/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:24 crc kubenswrapper[4727]: I0929 11:21:24.447963 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_725e53c7-0270-451e-8bde-851f9a60a8ab/nova-api-log/0.log" Sep 29 11:21:24 crc kubenswrapper[4727]: I0929 11:21:24.584854 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_45ccefb3-e0b3-4484-af4b-29b81d687507/nova-cell0-conductor-conductor/0.log" Sep 29 11:21:24 crc kubenswrapper[4727]: I0929 11:21:24.598166 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_725e53c7-0270-451e-8bde-851f9a60a8ab/nova-api-api/0.log" Sep 29 11:21:24 crc kubenswrapper[4727]: I0929 11:21:24.938798 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2522d266-4b11-4a05-96fc-69d5847a63c0/nova-cell1-conductor-conductor/0.log" Sep 29 11:21:24 crc kubenswrapper[4727]: I0929 11:21:24.995440 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6267f772-555e-4c6f-b937-6d8a4c61dab3/nova-cell1-novncproxy-novncproxy/0.log" Sep 29 11:21:25 crc kubenswrapper[4727]: I0929 11:21:25.229150 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-b78rw_a3618996-7c1c-480d-b868-25e17445d8a0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:25 crc kubenswrapper[4727]: I0929 11:21:25.510200 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f944aa92-5fe8-4100-9faf-00ef0443d80f/nova-metadata-log/0.log" Sep 29 11:21:25 crc kubenswrapper[4727]: I0929 11:21:25.962213 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_41ff8f24-0d46-49e0-88be-7b4d3c33d619/nova-scheduler-scheduler/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.098571 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/mysql-bootstrap/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.275458 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/mysql-bootstrap/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.325056 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/galera/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.657826 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/mysql-bootstrap/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.889739 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/galera/0.log" Sep 29 11:21:26 crc kubenswrapper[4727]: I0929 11:21:26.930400 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/mysql-bootstrap/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.082318 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f944aa92-5fe8-4100-9faf-00ef0443d80f/nova-metadata-metadata/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.203983 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_60736cbf-afd3-410e-87b8-bcf852ff2221/openstackclient/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.431618 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rvsjj_ad7a1be1-e75b-47f3-8b78-48f30fa37940/openstack-network-exporter/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.578452 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server-init/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.826765 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovs-vswitchd/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.902820 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server-init/0.log" Sep 29 11:21:27 crc kubenswrapper[4727]: I0929 11:21:27.907483 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.167952 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x86c4_36e70287-3b92-41e6-a056-fb29d1e03772/ovn-controller/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.370084 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-l9gxq_1dc57566-f99c-4041-a617-199b93c4e4eb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.509099 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2db0e30f-219c-4539-8f02-3f64f9f23a50/openstack-network-exporter/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.623271 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2db0e30f-219c-4539-8f02-3f64f9f23a50/ovn-northd/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.726221 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4bb550f2-83cb-472e-9e1e-b5da4779b42d/openstack-network-exporter/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.889025 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4bb550f2-83cb-472e-9e1e-b5da4779b42d/ovsdbserver-nb/0.log" Sep 29 11:21:28 crc kubenswrapper[4727]: I0929 11:21:28.959874 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13550429-27b2-4d79-8a60-447c329c77f6/openstack-network-exporter/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.177960 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13550429-27b2-4d79-8a60-447c329c77f6/ovsdbserver-sb/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.377014 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c44b4b898-64gpw_1b18f96d-e3b8-4758-8f71-b75a40cc9fbe/placement-api/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.553156 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c44b4b898-64gpw_1b18f96d-e3b8-4758-8f71-b75a40cc9fbe/placement-log/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.595540 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/setup-container/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.843536 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/setup-container/0.log" Sep 29 11:21:29 crc kubenswrapper[4727]: I0929 11:21:29.876537 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/rabbitmq/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.064384 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/setup-container/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.356712 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/setup-container/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.372799 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/rabbitmq/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.577133 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx_b6b1b4ed-2c0b-41f3-a542-e6fefaccc416/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.617290 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zbjd8_eac8b453-a168-458e-9e0b-28909a6323c4/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:30 crc kubenswrapper[4727]: I0929 11:21:30.932689 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j_874119b7-b7ce-4e87-a254-892df476efb0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.175014 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tfmmh_bc9712d6-5bbe-4749-acc0-a252a76a0ada/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.317866 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vx2sv_28a41bdb-b145-481a-9835-e08ce95e2e17/ssh-known-hosts-edpm-deployment/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.617200 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-598db7f995-m4hgm_87ecfca9-38cc-4b74-88b7-2d56d8f5638d/proxy-server/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.624243 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-598db7f995-m4hgm_87ecfca9-38cc-4b74-88b7-2d56d8f5638d/proxy-httpd/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.884255 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6jpg2_83f5e1ea-5026-48dc-a5a7-d6d841b45376/swift-ring-rebalance/0.log" Sep 29 11:21:31 crc kubenswrapper[4727]: I0929 11:21:31.987727 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-auditor/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.110687 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-reaper/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.118390 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-replicator/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.208098 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-server/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.322285 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-auditor/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.332216 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-replicator/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.434417 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-server/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.528853 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-updater/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.581492 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-auditor/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.644307 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-expirer/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.792022 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-server/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.796151 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-replicator/0.log" Sep 29 11:21:32 crc kubenswrapper[4727]: I0929 11:21:32.862238 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-updater/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.023616 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/rsync/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.048128 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/swift-recon-cron/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.342051 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw_77fc39d5-f092-4f06-9a6b-5d156935de57/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.373726 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0e997369-3d49-4bc4-95ec-77c503cda426/tempest-tests-tempest-tests-runner/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.545735 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_0518c511-9f69-42dd-88e3-661f13f8f019/test-operator-logs-container/0.log" Sep 29 11:21:33 crc kubenswrapper[4727]: I0929 11:21:33.760282 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c_a2a42153-5fde-492b-9564-fded49a50ebd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:21:41 crc kubenswrapper[4727]: I0929 11:21:41.334814 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8f2d7f61-fe27-4af2-8ef4-05d247fde09d/memcached/0.log" Sep 29 11:21:49 crc kubenswrapper[4727]: I0929 11:21:49.246889 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:21:49 crc kubenswrapper[4727]: I0929 11:21:49.247485 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.247001 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.248606 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.248735 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.250011 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.250089 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" gracePeriod=600 Sep 29 11:22:19 crc kubenswrapper[4727]: E0929 11:22:19.382970 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.813503 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" exitCode=0 Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.813542 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384"} Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.813601 4727 scope.go:117] "RemoveContainer" containerID="3e17f387f4f0e585b8000fb2a918d04e93e0b0f00b7b5d2088a849bf16d94c5f" Sep 29 11:22:19 crc kubenswrapper[4727]: I0929 11:22:19.814327 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:22:19 crc kubenswrapper[4727]: E0929 11:22:19.814731 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:22:28 crc kubenswrapper[4727]: I0929 11:22:28.927530 4727 generic.go:334] "Generic (PLEG): container finished" podID="208c21a2-8735-43ea-9396-b35f22696f66" containerID="82e40f9791267d39a35763c93a41de667a721f0089bd428df3a540f2011ae3f8" exitCode=0 Sep 29 11:22:28 crc kubenswrapper[4727]: I0929 11:22:28.927588 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" event={"ID":"208c21a2-8735-43ea-9396-b35f22696f66","Type":"ContainerDied","Data":"82e40f9791267d39a35763c93a41de667a721f0089bd428df3a540f2011ae3f8"} Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.038840 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.070019 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxc2m\" (UniqueName: \"kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m\") pod \"208c21a2-8735-43ea-9396-b35f22696f66\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.070237 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host\") pod \"208c21a2-8735-43ea-9396-b35f22696f66\" (UID: \"208c21a2-8735-43ea-9396-b35f22696f66\") " Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.070349 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host" (OuterVolumeSpecName: "host") pod "208c21a2-8735-43ea-9396-b35f22696f66" (UID: "208c21a2-8735-43ea-9396-b35f22696f66"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.070809 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c21a2-8735-43ea-9396-b35f22696f66-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.076047 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-jnzrt"] Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.081475 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m" (OuterVolumeSpecName: "kube-api-access-zxc2m") pod "208c21a2-8735-43ea-9396-b35f22696f66" (UID: "208c21a2-8735-43ea-9396-b35f22696f66"). InnerVolumeSpecName "kube-api-access-zxc2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.085661 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-jnzrt"] Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.172912 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxc2m\" (UniqueName: \"kubernetes.io/projected/208c21a2-8735-43ea-9396-b35f22696f66-kube-api-access-zxc2m\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.946392 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="582fd956a8dbf149daa4e15061893dbf65210203c40a0398683882987d376d95" Sep 29 11:22:30 crc kubenswrapper[4727]: I0929 11:22:30.946451 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-jnzrt" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.121935 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208c21a2-8735-43ea-9396-b35f22696f66" path="/var/lib/kubelet/pods/208c21a2-8735-43ea-9396-b35f22696f66/volumes" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.250981 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-ppwt2"] Sep 29 11:22:31 crc kubenswrapper[4727]: E0929 11:22:31.251521 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208c21a2-8735-43ea-9396-b35f22696f66" containerName="container-00" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.251542 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="208c21a2-8735-43ea-9396-b35f22696f66" containerName="container-00" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.251815 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="208c21a2-8735-43ea-9396-b35f22696f66" containerName="container-00" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.252484 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.293796 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.294214 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gvm\" (UniqueName: \"kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.396562 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.396629 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gvm\" (UniqueName: \"kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.396750 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.414783 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gvm\" (UniqueName: \"kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm\") pod \"crc-debug-ppwt2\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.573783 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:31 crc kubenswrapper[4727]: I0929 11:22:31.956432 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" event={"ID":"a1d6cf82-a764-485e-9543-8dd97d837404","Type":"ContainerStarted","Data":"2227b5bd188146c541abd799fbbb23499a068f479c70e7ff6e20b7e1536cf467"} Sep 29 11:22:32 crc kubenswrapper[4727]: I0929 11:22:32.969881 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" event={"ID":"a1d6cf82-a764-485e-9543-8dd97d837404","Type":"ContainerStarted","Data":"d82b3e06770a823dfed883fc8a0a412e53bd8d1f81ec6e21554c280db9e52cc8"} Sep 29 11:22:32 crc kubenswrapper[4727]: I0929 11:22:32.989714 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" podStartSLOduration=1.989697569 podStartE2EDuration="1.989697569s" podCreationTimestamp="2025-09-29 11:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:22:32.986863796 +0000 UTC m=+3623.160177158" watchObservedRunningTime="2025-09-29 11:22:32.989697569 +0000 UTC m=+3623.163010931" Sep 29 11:22:33 crc kubenswrapper[4727]: I0929 11:22:33.981859 4727 generic.go:334] "Generic (PLEG): container finished" podID="a1d6cf82-a764-485e-9543-8dd97d837404" containerID="d82b3e06770a823dfed883fc8a0a412e53bd8d1f81ec6e21554c280db9e52cc8" exitCode=0 Sep 29 11:22:33 crc kubenswrapper[4727]: I0929 11:22:33.981913 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" event={"ID":"a1d6cf82-a764-485e-9543-8dd97d837404","Type":"ContainerDied","Data":"d82b3e06770a823dfed883fc8a0a412e53bd8d1f81ec6e21554c280db9e52cc8"} Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.108127 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:22:34 crc kubenswrapper[4727]: E0929 11:22:34.108463 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.900315 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.903858 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.910673 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.954123 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.954477 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn978\" (UniqueName: \"kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:34 crc kubenswrapper[4727]: I0929 11:22:34.954730 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.057073 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.057169 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn978\" (UniqueName: \"kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.057242 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.057853 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.057990 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.079177 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn978\" (UniqueName: \"kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978\") pod \"redhat-marketplace-2q7pp\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.162299 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.242293 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.263836 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46gvm\" (UniqueName: \"kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm\") pod \"a1d6cf82-a764-485e-9543-8dd97d837404\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.263960 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host\") pod \"a1d6cf82-a764-485e-9543-8dd97d837404\" (UID: \"a1d6cf82-a764-485e-9543-8dd97d837404\") " Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.264485 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host" (OuterVolumeSpecName: "host") pod "a1d6cf82-a764-485e-9543-8dd97d837404" (UID: "a1d6cf82-a764-485e-9543-8dd97d837404"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.269661 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm" (OuterVolumeSpecName: "kube-api-access-46gvm") pod "a1d6cf82-a764-485e-9543-8dd97d837404" (UID: "a1d6cf82-a764-485e-9543-8dd97d837404"). InnerVolumeSpecName "kube-api-access-46gvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.366222 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46gvm\" (UniqueName: \"kubernetes.io/projected/a1d6cf82-a764-485e-9543-8dd97d837404-kube-api-access-46gvm\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.366267 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d6cf82-a764-485e-9543-8dd97d837404-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:35 crc kubenswrapper[4727]: I0929 11:22:35.722893 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:36 crc kubenswrapper[4727]: I0929 11:22:36.021451 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerStarted","Data":"ad6978088ee438f7c43609b1e72f2421a96deefbafa6411ed8b653011c3dbb39"} Sep 29 11:22:36 crc kubenswrapper[4727]: I0929 11:22:36.023296 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" event={"ID":"a1d6cf82-a764-485e-9543-8dd97d837404","Type":"ContainerDied","Data":"2227b5bd188146c541abd799fbbb23499a068f479c70e7ff6e20b7e1536cf467"} Sep 29 11:22:36 crc kubenswrapper[4727]: I0929 11:22:36.023322 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2227b5bd188146c541abd799fbbb23499a068f479c70e7ff6e20b7e1536cf467" Sep 29 11:22:36 crc kubenswrapper[4727]: I0929 11:22:36.023414 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-ppwt2" Sep 29 11:22:37 crc kubenswrapper[4727]: I0929 11:22:37.033510 4727 generic.go:334] "Generic (PLEG): container finished" podID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerID="14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349" exitCode=0 Sep 29 11:22:37 crc kubenswrapper[4727]: I0929 11:22:37.034680 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerDied","Data":"14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349"} Sep 29 11:22:39 crc kubenswrapper[4727]: I0929 11:22:39.077568 4727 generic.go:334] "Generic (PLEG): container finished" podID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerID="d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd" exitCode=0 Sep 29 11:22:39 crc kubenswrapper[4727]: I0929 11:22:39.077663 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerDied","Data":"d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd"} Sep 29 11:22:39 crc kubenswrapper[4727]: I0929 11:22:39.613330 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-ppwt2"] Sep 29 11:22:39 crc kubenswrapper[4727]: I0929 11:22:39.620715 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-ppwt2"] Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.777539 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-6dt84"] Sep 29 11:22:40 crc kubenswrapper[4727]: E0929 11:22:40.778262 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d6cf82-a764-485e-9543-8dd97d837404" containerName="container-00" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.778276 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d6cf82-a764-485e-9543-8dd97d837404" containerName="container-00" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.778546 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d6cf82-a764-485e-9543-8dd97d837404" containerName="container-00" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.779318 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.864598 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.864767 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5cvr\" (UniqueName: \"kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.966848 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.966978 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5cvr\" (UniqueName: \"kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.966980 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:40 crc kubenswrapper[4727]: I0929 11:22:40.985118 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5cvr\" (UniqueName: \"kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr\") pod \"crc-debug-6dt84\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:41 crc kubenswrapper[4727]: I0929 11:22:41.148602 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d6cf82-a764-485e-9543-8dd97d837404" path="/var/lib/kubelet/pods/a1d6cf82-a764-485e-9543-8dd97d837404/volumes" Sep 29 11:22:41 crc kubenswrapper[4727]: I0929 11:22:41.158756 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerStarted","Data":"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4"} Sep 29 11:22:41 crc kubenswrapper[4727]: I0929 11:22:41.163787 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.152745 4727 generic.go:334] "Generic (PLEG): container finished" podID="9f54e3cd-5266-489e-ace9-0cf22b68d253" containerID="03fa9552318c88f7bd27c6dcdfefacbef7bfdf808767466722d633dcc7a9cf7f" exitCode=0 Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.152824 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" event={"ID":"9f54e3cd-5266-489e-ace9-0cf22b68d253","Type":"ContainerDied","Data":"03fa9552318c88f7bd27c6dcdfefacbef7bfdf808767466722d633dcc7a9cf7f"} Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.153072 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" event={"ID":"9f54e3cd-5266-489e-ace9-0cf22b68d253","Type":"ContainerStarted","Data":"554bfeef9c4fb2805b5ced49c308a109705de772b81d573a130b33069442c1e5"} Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.172897 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2q7pp" podStartSLOduration=4.481467033 podStartE2EDuration="8.172881684s" podCreationTimestamp="2025-09-29 11:22:34 +0000 UTC" firstStartedPulling="2025-09-29 11:22:37.035728062 +0000 UTC m=+3627.209041424" lastFinishedPulling="2025-09-29 11:22:40.727142713 +0000 UTC m=+3630.900456075" observedRunningTime="2025-09-29 11:22:41.186790908 +0000 UTC m=+3631.360104270" watchObservedRunningTime="2025-09-29 11:22:42.172881684 +0000 UTC m=+3632.346195046" Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.191883 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-6dt84"] Sep 29 11:22:42 crc kubenswrapper[4727]: I0929 11:22:42.199288 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6h5r/crc-debug-6dt84"] Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.261419 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.307882 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host\") pod \"9f54e3cd-5266-489e-ace9-0cf22b68d253\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.307994 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5cvr\" (UniqueName: \"kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr\") pod \"9f54e3cd-5266-489e-ace9-0cf22b68d253\" (UID: \"9f54e3cd-5266-489e-ace9-0cf22b68d253\") " Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.308006 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host" (OuterVolumeSpecName: "host") pod "9f54e3cd-5266-489e-ace9-0cf22b68d253" (UID: "9f54e3cd-5266-489e-ace9-0cf22b68d253"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.308453 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f54e3cd-5266-489e-ace9-0cf22b68d253-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.313969 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr" (OuterVolumeSpecName: "kube-api-access-x5cvr") pod "9f54e3cd-5266-489e-ace9-0cf22b68d253" (UID: "9f54e3cd-5266-489e-ace9-0cf22b68d253"). InnerVolumeSpecName "kube-api-access-x5cvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.410418 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5cvr\" (UniqueName: \"kubernetes.io/projected/9f54e3cd-5266-489e-ace9-0cf22b68d253-kube-api-access-x5cvr\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.676385 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.834486 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.857719 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:22:43 crc kubenswrapper[4727]: I0929 11:22:43.896845 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.062574 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/extract/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.105687 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.145417 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.170406 4727 scope.go:117] "RemoveContainer" containerID="03fa9552318c88f7bd27c6dcdfefacbef7bfdf808767466722d633dcc7a9cf7f" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.170436 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/crc-debug-6dt84" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.252878 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6495d75b5-rbtdb_44a5b923-82b5-4af1-ad00-ba65998598ff/kube-rbac-proxy/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.370058 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6495d75b5-rbtdb_44a5b923-82b5-4af1-ad00-ba65998598ff/manager/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.470637 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-748c574d75-8s5z5_f6f41603-6f22-4086-be87-c1a0062a691d/kube-rbac-proxy/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.473611 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-748c574d75-8s5z5_f6f41603-6f22-4086-be87-c1a0062a691d/manager/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.593646 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d74f4d695-8txnk_2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609/kube-rbac-proxy/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.661384 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d74f4d695-8txnk_2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609/manager/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.786077 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67b5d44b7f-vx2tw_2cb520ac-7d2f-46ab-b8bc-5f515485d278/kube-rbac-proxy/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.841439 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67b5d44b7f-vx2tw_2cb520ac-7d2f-46ab-b8bc-5f515485d278/manager/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.858710 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8ff95898-6b6gt_aef52f91-e131-49a0-84c8-ba76db16d118/kube-rbac-proxy/0.log" Sep 29 11:22:44 crc kubenswrapper[4727]: I0929 11:22:44.973116 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8ff95898-6b6gt_aef52f91-e131-49a0-84c8-ba76db16d118/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.046462 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-695847bc78-fcs2w_66342226-bc85-46a0-8d84-0627c26b196c/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.077028 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-695847bc78-fcs2w_66342226-bc85-46a0-8d84-0627c26b196c/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.121192 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f54e3cd-5266-489e-ace9-0cf22b68d253" path="/var/lib/kubelet/pods/9f54e3cd-5266-489e-ace9-0cf22b68d253/volumes" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.242515 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.242564 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.261958 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858cd69f49-swh7z_e5545a7b-7c61-4e6b-92e9-3a22045a2133/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.306029 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.390181 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9fc8d5567-h5bnt_f4fac1bc-f504-4b82-8371-1c7afd6f76c7/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.391974 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858cd69f49-swh7z_e5545a7b-7c61-4e6b-92e9-3a22045a2133/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.448324 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9fc8d5567-h5bnt_f4fac1bc-f504-4b82-8371-1c7afd6f76c7/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.560316 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7bf498966c-vktwk_6f3ebbd5-f00d-4db3-8160-03e9dce230e4/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.653766 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7bf498966c-vktwk_6f3ebbd5-f00d-4db3-8160-03e9dce230e4/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.702088 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56cf9c6b99-5xzsf_51327d2c-111c-4201-9f5d-ec9bf7f1d31f/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.762862 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56cf9c6b99-5xzsf_51327d2c-111c-4201-9f5d-ec9bf7f1d31f/manager/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.872545 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-687b9cf756-z5zqv_f0ad5209-9d60-488c-9cb7-3252bb844305/kube-rbac-proxy/0.log" Sep 29 11:22:45 crc kubenswrapper[4727]: I0929 11:22:45.899541 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-687b9cf756-z5zqv_f0ad5209-9d60-488c-9cb7-3252bb844305/manager/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.037323 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54d766c9f9-jxvz6_ff90604b-0b99-4c43-b38a-d467d74a6376/kube-rbac-proxy/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.091112 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54d766c9f9-jxvz6_ff90604b-0b99-4c43-b38a-d467d74a6376/manager/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.108872 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:22:46 crc kubenswrapper[4727]: E0929 11:22:46.109301 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.159544 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-whv4l_3164f156-e48b-4eaa-8e29-ecf414254d3a/kube-rbac-proxy/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.278112 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.339109 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.345468 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-47bvr_495ff530-82ab-48b5-9c58-1f03614f5b5a/kube-rbac-proxy/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.347626 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-whv4l_3164f156-e48b-4eaa-8e29-ecf414254d3a/manager/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.434265 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-47bvr_495ff530-82ab-48b5-9c58-1f03614f5b5a/manager/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.520397 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-sp6px_a6dfa756-7117-4d5c-830b-a3dc23d86682/kube-rbac-proxy/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.541262 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-sp6px_a6dfa756-7117-4d5c-830b-a3dc23d86682/manager/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.722963 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-578979c99-jgw9d_ca7d374b-4f00-4831-b2b1-39fbf673b820/kube-rbac-proxy/0.log" Sep 29 11:22:46 crc kubenswrapper[4727]: I0929 11:22:46.983123 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bb87788c-jvd62_b2fb3066-6094-4799-b68d-67ab396b160e/kube-rbac-proxy/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.155944 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bb87788c-jvd62_b2fb3066-6094-4799-b68d-67ab396b160e/operator/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.203292 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ps6m_9d7c10ba-9139-47d6-a681-507681066bff/registry-server/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.363104 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5f95c46c78-6rsg9_19c9fecd-d2f6-453c-be12-ede6f08fcf0f/kube-rbac-proxy/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.449462 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5f95c46c78-6rsg9_19c9fecd-d2f6-453c-be12-ede6f08fcf0f/manager/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.539605 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-774b97b48-j7hz6_9d628165-af26-41a6-b05c-e633673213ab/kube-rbac-proxy/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.635846 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-774b97b48-j7hz6_9d628165-af26-41a6-b05c-e633673213ab/manager/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.764786 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-dgghl_120b089a-5999-450b-acd7-5f48a5a105bd/operator/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.779928 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-578979c99-jgw9d_ca7d374b-4f00-4831-b2b1-39fbf673b820/manager/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.835673 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-dsg68_7c7f0180-cbee-4a25-b069-3b11a90b2d08/kube-rbac-proxy/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.905346 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-dsg68_7c7f0180-cbee-4a25-b069-3b11a90b2d08/manager/0.log" Sep 29 11:22:47 crc kubenswrapper[4727]: I0929 11:22:47.940357 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5bf96cfbc4-zstln_a64b04d3-a9a3-4c5e-a356-045f7b13c572/kube-rbac-proxy/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.094602 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5bf96cfbc4-zstln_a64b04d3-a9a3-4c5e-a356-045f7b13c572/manager/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.128991 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6p49k_baad2063-56c9-414d-b8da-5969bf4f6220/kube-rbac-proxy/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.129514 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6p49k_baad2063-56c9-414d-b8da-5969bf4f6220/manager/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.218407 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2q7pp" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="registry-server" containerID="cri-o://005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4" gracePeriod=2 Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.246175 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-djbkf_fadeace3-e8ba-4e11-a2f7-6dee11f875df/kube-rbac-proxy/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.283419 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-djbkf_fadeace3-e8ba-4e11-a2f7-6dee11f875df/manager/0.log" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.698893 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.849477 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities\") pod \"b4081b10-7bb1-4238-b187-0b5f4e436222\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.849570 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content\") pod \"b4081b10-7bb1-4238-b187-0b5f4e436222\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.849670 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn978\" (UniqueName: \"kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978\") pod \"b4081b10-7bb1-4238-b187-0b5f4e436222\" (UID: \"b4081b10-7bb1-4238-b187-0b5f4e436222\") " Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.850697 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities" (OuterVolumeSpecName: "utilities") pod "b4081b10-7bb1-4238-b187-0b5f4e436222" (UID: "b4081b10-7bb1-4238-b187-0b5f4e436222"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.851193 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.856045 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978" (OuterVolumeSpecName: "kube-api-access-xn978") pod "b4081b10-7bb1-4238-b187-0b5f4e436222" (UID: "b4081b10-7bb1-4238-b187-0b5f4e436222"). InnerVolumeSpecName "kube-api-access-xn978". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.866839 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4081b10-7bb1-4238-b187-0b5f4e436222" (UID: "b4081b10-7bb1-4238-b187-0b5f4e436222"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.952759 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn978\" (UniqueName: \"kubernetes.io/projected/b4081b10-7bb1-4238-b187-0b5f4e436222-kube-api-access-xn978\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:48 crc kubenswrapper[4727]: I0929 11:22:48.952998 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4081b10-7bb1-4238-b187-0b5f4e436222-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.229106 4727 generic.go:334] "Generic (PLEG): container finished" podID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerID="005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4" exitCode=0 Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.229159 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerDied","Data":"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4"} Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.229192 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q7pp" event={"ID":"b4081b10-7bb1-4238-b187-0b5f4e436222","Type":"ContainerDied","Data":"ad6978088ee438f7c43609b1e72f2421a96deefbafa6411ed8b653011c3dbb39"} Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.229213 4727 scope.go:117] "RemoveContainer" containerID="005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.229423 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q7pp" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.252814 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.253585 4727 scope.go:117] "RemoveContainer" containerID="d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.264657 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q7pp"] Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.277379 4727 scope.go:117] "RemoveContainer" containerID="14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.331695 4727 scope.go:117] "RemoveContainer" containerID="005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4" Sep 29 11:22:49 crc kubenswrapper[4727]: E0929 11:22:49.332178 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4\": container with ID starting with 005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4 not found: ID does not exist" containerID="005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.332218 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4"} err="failed to get container status \"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4\": rpc error: code = NotFound desc = could not find container \"005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4\": container with ID starting with 005370f57846dc3a8158d4c49e1d90b0f5e4e0488c04e21c0f0d910f392ca4f4 not found: ID does not exist" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.332245 4727 scope.go:117] "RemoveContainer" containerID="d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd" Sep 29 11:22:49 crc kubenswrapper[4727]: E0929 11:22:49.332724 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd\": container with ID starting with d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd not found: ID does not exist" containerID="d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.332755 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd"} err="failed to get container status \"d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd\": rpc error: code = NotFound desc = could not find container \"d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd\": container with ID starting with d6764f150602d93915de30f87b2ff8620d8250fc25e420da70a078d8a9cfe4bd not found: ID does not exist" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.332783 4727 scope.go:117] "RemoveContainer" containerID="14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349" Sep 29 11:22:49 crc kubenswrapper[4727]: E0929 11:22:49.333007 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349\": container with ID starting with 14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349 not found: ID does not exist" containerID="14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349" Sep 29 11:22:49 crc kubenswrapper[4727]: I0929 11:22:49.333024 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349"} err="failed to get container status \"14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349\": rpc error: code = NotFound desc = could not find container \"14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349\": container with ID starting with 14751b17a926052400ceaedabefa4888b6dc420b2a38351a2811eeafbb1d7349 not found: ID does not exist" Sep 29 11:22:51 crc kubenswrapper[4727]: I0929 11:22:51.118857 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" path="/var/lib/kubelet/pods/b4081b10-7bb1-4238-b187-0b5f4e436222/volumes" Sep 29 11:23:01 crc kubenswrapper[4727]: I0929 11:23:01.114248 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:01 crc kubenswrapper[4727]: E0929 11:23:01.115260 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:23:02 crc kubenswrapper[4727]: I0929 11:23:02.443303 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-f7hn5_46124f8a-3360-4dfe-b468-5addce47a614/control-plane-machine-set-operator/0.log" Sep 29 11:23:02 crc kubenswrapper[4727]: I0929 11:23:02.598917 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lw586_6404327b-095b-435a-b953-b1c3b7a6f332/kube-rbac-proxy/0.log" Sep 29 11:23:02 crc kubenswrapper[4727]: I0929 11:23:02.640161 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lw586_6404327b-095b-435a-b953-b1c3b7a6f332/machine-api-operator/0.log" Sep 29 11:23:12 crc kubenswrapper[4727]: I0929 11:23:12.108921 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:12 crc kubenswrapper[4727]: E0929 11:23:12.109800 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:23:13 crc kubenswrapper[4727]: I0929 11:23:13.321090 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wxxch_aec4835b-f46f-4b6a-aab2-37154553a6aa/cert-manager-controller/0.log" Sep 29 11:23:13 crc kubenswrapper[4727]: I0929 11:23:13.370308 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-62b6k_15e48513-d028-479d-9796-ab2ecbe72f58/cert-manager-cainjector/0.log" Sep 29 11:23:13 crc kubenswrapper[4727]: I0929 11:23:13.508223 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-6w54j_4e4ebfd5-7776-4705-943f-b6ac0e6a6238/cert-manager-webhook/0.log" Sep 29 11:23:23 crc kubenswrapper[4727]: I0929 11:23:23.898846 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-l7xjv_8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6/nmstate-console-plugin/0.log" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.108724 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:24 crc kubenswrapper[4727]: E0929 11:23:24.109367 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.115648 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2hx48_a713c14e-bce4-44a3-940d-0a987cc6fa5b/kube-rbac-proxy/0.log" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.126075 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sh4hc_8109dd98-72c1-4542-a34c-1f5d911b5fd4/nmstate-handler/0.log" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.173147 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2hx48_a713c14e-bce4-44a3-940d-0a987cc6fa5b/nmstate-metrics/0.log" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.323855 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-5vcth_2ef4df09-2078-4ef8-9c7b-8e637b20cfcf/nmstate-operator/0.log" Sep 29 11:23:24 crc kubenswrapper[4727]: I0929 11:23:24.356731 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-shptq_7655e999-03ae-45f0-a123-1efa69626f7b/nmstate-webhook/0.log" Sep 29 11:23:35 crc kubenswrapper[4727]: I0929 11:23:35.109235 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:35 crc kubenswrapper[4727]: E0929 11:23:35.110080 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:23:36 crc kubenswrapper[4727]: I0929 11:23:36.606842 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-b8fqx_5a5b2bf7-f24e-494d-a0fc-11c6f90737bd/kube-rbac-proxy/0.log" Sep 29 11:23:36 crc kubenswrapper[4727]: I0929 11:23:36.781115 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-b8fqx_5a5b2bf7-f24e-494d-a0fc-11c6f90737bd/controller/0.log" Sep 29 11:23:36 crc kubenswrapper[4727]: I0929 11:23:36.843836 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.012816 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.014522 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.035263 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.046292 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.194084 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.200570 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.212390 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.253248 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.431204 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.443021 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.466029 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.470036 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/controller/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.648991 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/kube-rbac-proxy/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.662605 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/kube-rbac-proxy-frr/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.664661 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/frr-metrics/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.833882 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/reloader/0.log" Sep 29 11:23:37 crc kubenswrapper[4727]: I0929 11:23:37.896464 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-6pmt2_424bbcbe-8ea1-491c-b1b0-55285bff680e/frr-k8s-webhook-server/0.log" Sep 29 11:23:38 crc kubenswrapper[4727]: I0929 11:23:38.123603 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-86747676bf-9fmjz_9e45a861-9d6a-43af-9235-33e4cea43955/manager/0.log" Sep 29 11:23:38 crc kubenswrapper[4727]: I0929 11:23:38.263713 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b8c57d74f-nnlj8_5c0a38f9-cb25-4466-9d31-4d6bf3943989/webhook-server/0.log" Sep 29 11:23:38 crc kubenswrapper[4727]: I0929 11:23:38.380960 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vh9j9_fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e/kube-rbac-proxy/0.log" Sep 29 11:23:38 crc kubenswrapper[4727]: I0929 11:23:38.919211 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vh9j9_fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e/speaker/0.log" Sep 29 11:23:39 crc kubenswrapper[4727]: I0929 11:23:39.072048 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/frr/0.log" Sep 29 11:23:46 crc kubenswrapper[4727]: I0929 11:23:46.108688 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:46 crc kubenswrapper[4727]: E0929 11:23:46.110442 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:23:50 crc kubenswrapper[4727]: I0929 11:23:50.869947 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.189021 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.221801 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.280756 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.454132 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.507105 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.541482 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/extract/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.633584 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.860457 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.893185 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:23:51 crc kubenswrapper[4727]: I0929 11:23:51.899886 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.109796 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.195560 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.327668 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.540506 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/registry-server/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.579537 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.617233 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.644763 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.809130 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:23:52 crc kubenswrapper[4727]: I0929 11:23:52.809141 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.041481 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.290826 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.323615 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.331067 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.550930 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.581789 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.590232 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/extract/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.681103 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/registry-server/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.821035 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nsbj_a9c2af09-6645-4cef-bba5-59c2b5b278d2/marketplace-operator/0.log" Sep 29 11:23:53 crc kubenswrapper[4727]: I0929 11:23:53.890987 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.120352 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.123571 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.171502 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.331921 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.331962 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.427778 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/registry-server/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.519506 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.680653 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.680653 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.735153 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.911440 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:23:54 crc kubenswrapper[4727]: I0929 11:23:54.926277 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:23:55 crc kubenswrapper[4727]: I0929 11:23:55.206442 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/registry-server/0.log" Sep 29 11:23:58 crc kubenswrapper[4727]: I0929 11:23:58.109166 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:23:58 crc kubenswrapper[4727]: E0929 11:23:58.109706 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:24:09 crc kubenswrapper[4727]: I0929 11:24:09.108746 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:24:09 crc kubenswrapper[4727]: E0929 11:24:09.109602 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:24:24 crc kubenswrapper[4727]: I0929 11:24:24.109664 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:24:24 crc kubenswrapper[4727]: E0929 11:24:24.110305 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:24:39 crc kubenswrapper[4727]: I0929 11:24:39.109107 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:24:39 crc kubenswrapper[4727]: E0929 11:24:39.109975 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:24:51 crc kubenswrapper[4727]: I0929 11:24:51.118016 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:24:51 crc kubenswrapper[4727]: E0929 11:24:51.118914 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:25:05 crc kubenswrapper[4727]: I0929 11:25:05.116884 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:25:05 crc kubenswrapper[4727]: E0929 11:25:05.117845 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:25:19 crc kubenswrapper[4727]: I0929 11:25:19.109245 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:25:19 crc kubenswrapper[4727]: E0929 11:25:19.110076 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:25:33 crc kubenswrapper[4727]: I0929 11:25:33.112928 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:25:33 crc kubenswrapper[4727]: E0929 11:25:33.113639 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.662812 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:40 crc kubenswrapper[4727]: E0929 11:25:40.664832 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="extract-content" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.664853 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="extract-content" Sep 29 11:25:40 crc kubenswrapper[4727]: E0929 11:25:40.664913 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="registry-server" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.664921 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="registry-server" Sep 29 11:25:40 crc kubenswrapper[4727]: E0929 11:25:40.664937 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="extract-utilities" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.664961 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="extract-utilities" Sep 29 11:25:40 crc kubenswrapper[4727]: E0929 11:25:40.664975 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f54e3cd-5266-489e-ace9-0cf22b68d253" containerName="container-00" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.664980 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f54e3cd-5266-489e-ace9-0cf22b68d253" containerName="container-00" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.665202 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4081b10-7bb1-4238-b187-0b5f4e436222" containerName="registry-server" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.665220 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f54e3cd-5266-489e-ace9-0cf22b68d253" containerName="container-00" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.666966 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.677597 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.818088 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4zr\" (UniqueName: \"kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.818201 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.818238 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.920647 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.921037 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.921196 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4zr\" (UniqueName: \"kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.921254 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.921619 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.944333 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4zr\" (UniqueName: \"kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr\") pod \"community-operators-smjs7\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:40 crc kubenswrapper[4727]: I0929 11:25:40.990516 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:41 crc kubenswrapper[4727]: I0929 11:25:41.512967 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:41 crc kubenswrapper[4727]: I0929 11:25:41.795875 4727 generic.go:334] "Generic (PLEG): container finished" podID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerID="5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07" exitCode=0 Sep 29 11:25:41 crc kubenswrapper[4727]: I0929 11:25:41.796008 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerDied","Data":"5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07"} Sep 29 11:25:41 crc kubenswrapper[4727]: I0929 11:25:41.796241 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerStarted","Data":"8ea4b3d8e5140f16992ad4626a65ad8c922a3795004f3cb92506d3847abf44f1"} Sep 29 11:25:41 crc kubenswrapper[4727]: I0929 11:25:41.798461 4727 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 11:25:42 crc kubenswrapper[4727]: I0929 11:25:42.809457 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerStarted","Data":"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01"} Sep 29 11:25:43 crc kubenswrapper[4727]: I0929 11:25:43.822967 4727 generic.go:334] "Generic (PLEG): container finished" podID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerID="12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01" exitCode=0 Sep 29 11:25:43 crc kubenswrapper[4727]: I0929 11:25:43.823047 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerDied","Data":"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01"} Sep 29 11:25:44 crc kubenswrapper[4727]: I0929 11:25:44.832931 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerStarted","Data":"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39"} Sep 29 11:25:46 crc kubenswrapper[4727]: I0929 11:25:46.108825 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:25:46 crc kubenswrapper[4727]: E0929 11:25:46.109114 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:25:50 crc kubenswrapper[4727]: I0929 11:25:50.991549 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:50 crc kubenswrapper[4727]: I0929 11:25:50.992889 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:51 crc kubenswrapper[4727]: I0929 11:25:51.034849 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:51 crc kubenswrapper[4727]: I0929 11:25:51.055802 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-smjs7" podStartSLOduration=8.527329977 podStartE2EDuration="11.055784351s" podCreationTimestamp="2025-09-29 11:25:40 +0000 UTC" firstStartedPulling="2025-09-29 11:25:41.798189222 +0000 UTC m=+3811.971502584" lastFinishedPulling="2025-09-29 11:25:44.326643596 +0000 UTC m=+3814.499956958" observedRunningTime="2025-09-29 11:25:44.860490932 +0000 UTC m=+3815.033804304" watchObservedRunningTime="2025-09-29 11:25:51.055784351 +0000 UTC m=+3821.229097703" Sep 29 11:25:51 crc kubenswrapper[4727]: I0929 11:25:51.937779 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:51 crc kubenswrapper[4727]: I0929 11:25:51.994848 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:53 crc kubenswrapper[4727]: I0929 11:25:53.916515 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-smjs7" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="registry-server" containerID="cri-o://b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39" gracePeriod=2 Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.328126 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.489178 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities\") pod \"44785ecb-485e-4c3f-8e8d-805bb4314d43\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.489268 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content\") pod \"44785ecb-485e-4c3f-8e8d-805bb4314d43\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.489445 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k4zr\" (UniqueName: \"kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr\") pod \"44785ecb-485e-4c3f-8e8d-805bb4314d43\" (UID: \"44785ecb-485e-4c3f-8e8d-805bb4314d43\") " Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.490209 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities" (OuterVolumeSpecName: "utilities") pod "44785ecb-485e-4c3f-8e8d-805bb4314d43" (UID: "44785ecb-485e-4c3f-8e8d-805bb4314d43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.494557 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr" (OuterVolumeSpecName: "kube-api-access-6k4zr") pod "44785ecb-485e-4c3f-8e8d-805bb4314d43" (UID: "44785ecb-485e-4c3f-8e8d-805bb4314d43"). InnerVolumeSpecName "kube-api-access-6k4zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.541323 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44785ecb-485e-4c3f-8e8d-805bb4314d43" (UID: "44785ecb-485e-4c3f-8e8d-805bb4314d43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.591046 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.591072 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44785ecb-485e-4c3f-8e8d-805bb4314d43-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.591084 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k4zr\" (UniqueName: \"kubernetes.io/projected/44785ecb-485e-4c3f-8e8d-805bb4314d43-kube-api-access-6k4zr\") on node \"crc\" DevicePath \"\"" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.930913 4727 generic.go:334] "Generic (PLEG): container finished" podID="efeb8425-b618-4251-ad73-3be72e98289b" containerID="b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b" exitCode=0 Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.930941 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" event={"ID":"efeb8425-b618-4251-ad73-3be72e98289b","Type":"ContainerDied","Data":"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b"} Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.931583 4727 scope.go:117] "RemoveContainer" containerID="b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.935953 4727 generic.go:334] "Generic (PLEG): container finished" podID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerID="b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39" exitCode=0 Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.935995 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerDied","Data":"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39"} Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.936025 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smjs7" event={"ID":"44785ecb-485e-4c3f-8e8d-805bb4314d43","Type":"ContainerDied","Data":"8ea4b3d8e5140f16992ad4626a65ad8c922a3795004f3cb92506d3847abf44f1"} Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.936046 4727 scope.go:117] "RemoveContainer" containerID="b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.936269 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smjs7" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.965507 4727 scope.go:117] "RemoveContainer" containerID="12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01" Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.977467 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.986584 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-smjs7"] Sep 29 11:25:54 crc kubenswrapper[4727]: I0929 11:25:54.990580 4727 scope.go:117] "RemoveContainer" containerID="5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.036135 4727 scope.go:117] "RemoveContainer" containerID="b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39" Sep 29 11:25:55 crc kubenswrapper[4727]: E0929 11:25:55.036688 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39\": container with ID starting with b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39 not found: ID does not exist" containerID="b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.036733 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39"} err="failed to get container status \"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39\": rpc error: code = NotFound desc = could not find container \"b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39\": container with ID starting with b5a48e647a9f840827b258dce190a9fad0f4eb460c79117e63ac529956cf0a39 not found: ID does not exist" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.036765 4727 scope.go:117] "RemoveContainer" containerID="12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01" Sep 29 11:25:55 crc kubenswrapper[4727]: E0929 11:25:55.037036 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01\": container with ID starting with 12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01 not found: ID does not exist" containerID="12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.037061 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01"} err="failed to get container status \"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01\": rpc error: code = NotFound desc = could not find container \"12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01\": container with ID starting with 12c151b9f25ec8d0ed68db464322763a39a64ca6766203214d9c10d143634e01 not found: ID does not exist" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.037080 4727 scope.go:117] "RemoveContainer" containerID="5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07" Sep 29 11:25:55 crc kubenswrapper[4727]: E0929 11:25:55.037285 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07\": container with ID starting with 5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07 not found: ID does not exist" containerID="5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.037311 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07"} err="failed to get container status \"5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07\": rpc error: code = NotFound desc = could not find container \"5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07\": container with ID starting with 5a50362c60c07f5c9f10911e574bc67c3acca25f43718ecbb02554ecfa24bf07 not found: ID does not exist" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.117965 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" path="/var/lib/kubelet/pods/44785ecb-485e-4c3f-8e8d-805bb4314d43/volumes" Sep 29 11:25:55 crc kubenswrapper[4727]: I0929 11:25:55.808321 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6h5r_must-gather-fh9v6_efeb8425-b618-4251-ad73-3be72e98289b/gather/0.log" Sep 29 11:26:01 crc kubenswrapper[4727]: I0929 11:26:01.125214 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:26:01 crc kubenswrapper[4727]: E0929 11:26:01.126426 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.207533 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6h5r/must-gather-fh9v6"] Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.208377 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="copy" containerID="cri-o://16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff" gracePeriod=2 Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.214841 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6h5r/must-gather-fh9v6"] Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.614721 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6h5r_must-gather-fh9v6_efeb8425-b618-4251-ad73-3be72e98289b/copy/0.log" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.615320 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.768180 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kql42\" (UniqueName: \"kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42\") pod \"efeb8425-b618-4251-ad73-3be72e98289b\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.768299 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output\") pod \"efeb8425-b618-4251-ad73-3be72e98289b\" (UID: \"efeb8425-b618-4251-ad73-3be72e98289b\") " Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.774690 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42" (OuterVolumeSpecName: "kube-api-access-kql42") pod "efeb8425-b618-4251-ad73-3be72e98289b" (UID: "efeb8425-b618-4251-ad73-3be72e98289b"). InnerVolumeSpecName "kube-api-access-kql42". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.870943 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kql42\" (UniqueName: \"kubernetes.io/projected/efeb8425-b618-4251-ad73-3be72e98289b-kube-api-access-kql42\") on node \"crc\" DevicePath \"\"" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.930595 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "efeb8425-b618-4251-ad73-3be72e98289b" (UID: "efeb8425-b618-4251-ad73-3be72e98289b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:26:04 crc kubenswrapper[4727]: I0929 11:26:04.973639 4727 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efeb8425-b618-4251-ad73-3be72e98289b-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.026178 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6h5r_must-gather-fh9v6_efeb8425-b618-4251-ad73-3be72e98289b/copy/0.log" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.026542 4727 generic.go:334] "Generic (PLEG): container finished" podID="efeb8425-b618-4251-ad73-3be72e98289b" containerID="16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff" exitCode=143 Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.026597 4727 scope.go:117] "RemoveContainer" containerID="16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.026605 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6h5r/must-gather-fh9v6" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.049051 4727 scope.go:117] "RemoveContainer" containerID="b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.123899 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efeb8425-b618-4251-ad73-3be72e98289b" path="/var/lib/kubelet/pods/efeb8425-b618-4251-ad73-3be72e98289b/volumes" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.131017 4727 scope.go:117] "RemoveContainer" containerID="16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff" Sep 29 11:26:05 crc kubenswrapper[4727]: E0929 11:26:05.131640 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff\": container with ID starting with 16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff not found: ID does not exist" containerID="16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.131680 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff"} err="failed to get container status \"16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff\": rpc error: code = NotFound desc = could not find container \"16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff\": container with ID starting with 16b156c46710a0f9976f460511dfc00ff837de02b305642019d044a1784ee6ff not found: ID does not exist" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.131710 4727 scope.go:117] "RemoveContainer" containerID="b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b" Sep 29 11:26:05 crc kubenswrapper[4727]: E0929 11:26:05.132083 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b\": container with ID starting with b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b not found: ID does not exist" containerID="b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b" Sep 29 11:26:05 crc kubenswrapper[4727]: I0929 11:26:05.132166 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b"} err="failed to get container status \"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b\": rpc error: code = NotFound desc = could not find container \"b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b\": container with ID starting with b86ec562647850635a2ffbd3c3bec1ab672c3a743cbbb9ba90b7377bd0cb007b not found: ID does not exist" Sep 29 11:26:15 crc kubenswrapper[4727]: I0929 11:26:15.108007 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:26:15 crc kubenswrapper[4727]: E0929 11:26:15.108852 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:26:29 crc kubenswrapper[4727]: I0929 11:26:29.108908 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:26:29 crc kubenswrapper[4727]: E0929 11:26:29.109631 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.875205 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hfjd5/must-gather-ghqtn"] Sep 29 11:26:34 crc kubenswrapper[4727]: E0929 11:26:34.877365 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="extract-utilities" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.877390 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="extract-utilities" Sep 29 11:26:34 crc kubenswrapper[4727]: E0929 11:26:34.877429 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="gather" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.877442 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="gather" Sep 29 11:26:34 crc kubenswrapper[4727]: E0929 11:26:34.877503 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="registry-server" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.877517 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="registry-server" Sep 29 11:26:34 crc kubenswrapper[4727]: E0929 11:26:34.877568 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="copy" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.877580 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="copy" Sep 29 11:26:34 crc kubenswrapper[4727]: E0929 11:26:34.877610 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="extract-content" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.877620 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="extract-content" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.879324 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="44785ecb-485e-4c3f-8e8d-805bb4314d43" containerName="registry-server" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.879392 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="gather" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.879414 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="efeb8425-b618-4251-ad73-3be72e98289b" containerName="copy" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.895992 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.896040 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hfjd5/must-gather-ghqtn"] Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.900152 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hfjd5"/"kube-root-ca.crt" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.902207 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hfjd5"/"default-dockercfg-zw7r7" Sep 29 11:26:34 crc kubenswrapper[4727]: I0929 11:26:34.902494 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hfjd5"/"openshift-service-ca.crt" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.055891 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx4p5\" (UniqueName: \"kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.055988 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.157138 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx4p5\" (UniqueName: \"kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.157184 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.157954 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.176423 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx4p5\" (UniqueName: \"kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5\") pod \"must-gather-ghqtn\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.242100 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:26:35 crc kubenswrapper[4727]: I0929 11:26:35.759430 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hfjd5/must-gather-ghqtn"] Sep 29 11:26:36 crc kubenswrapper[4727]: I0929 11:26:36.298328 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" event={"ID":"79746d22-d0ae-40d7-a444-4eb4e162e75d","Type":"ContainerStarted","Data":"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760"} Sep 29 11:26:36 crc kubenswrapper[4727]: I0929 11:26:36.298738 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" event={"ID":"79746d22-d0ae-40d7-a444-4eb4e162e75d","Type":"ContainerStarted","Data":"398bb80cf6670dd0cb5e68d529086a05dda04191169684e1d6602a80f21cf418"} Sep 29 11:26:37 crc kubenswrapper[4727]: I0929 11:26:37.317077 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" event={"ID":"79746d22-d0ae-40d7-a444-4eb4e162e75d","Type":"ContainerStarted","Data":"aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1"} Sep 29 11:26:37 crc kubenswrapper[4727]: I0929 11:26:37.337193 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" podStartSLOduration=3.337173885 podStartE2EDuration="3.337173885s" podCreationTimestamp="2025-09-29 11:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:26:37.332018085 +0000 UTC m=+3867.505331447" watchObservedRunningTime="2025-09-29 11:26:37.337173885 +0000 UTC m=+3867.510487247" Sep 29 11:26:38 crc kubenswrapper[4727]: E0929 11:26:38.557485 4727 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.153:51950->38.102.83.153:44021: write tcp 38.102.83.153:51950->38.102.83.153:44021: write: broken pipe Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.501451 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-nmtn7"] Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.502551 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.601289 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvzhd\" (UniqueName: \"kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.602220 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.706622 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvzhd\" (UniqueName: \"kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.706730 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.706962 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.748283 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvzhd\" (UniqueName: \"kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd\") pod \"crc-debug-nmtn7\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:39 crc kubenswrapper[4727]: I0929 11:26:39.819902 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:26:40 crc kubenswrapper[4727]: I0929 11:26:40.351026 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" event={"ID":"79a50198-2083-42d0-bde4-3879b2fb4ec6","Type":"ContainerStarted","Data":"40940ba9d8707d663eb3a397c737eb8a71728dbbbb910be5da2350828e4a9b5a"} Sep 29 11:26:40 crc kubenswrapper[4727]: I0929 11:26:40.351637 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" event={"ID":"79a50198-2083-42d0-bde4-3879b2fb4ec6","Type":"ContainerStarted","Data":"f45e4b7ec9a0a2fdf05cfdab7c23559223306682e41e90e85483c42a169d5a0d"} Sep 29 11:26:40 crc kubenswrapper[4727]: I0929 11:26:40.371483 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" podStartSLOduration=1.371464913 podStartE2EDuration="1.371464913s" podCreationTimestamp="2025-09-29 11:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:26:40.366028547 +0000 UTC m=+3870.539341909" watchObservedRunningTime="2025-09-29 11:26:40.371464913 +0000 UTC m=+3870.544778275" Sep 29 11:26:42 crc kubenswrapper[4727]: I0929 11:26:42.108447 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:26:42 crc kubenswrapper[4727]: E0929 11:26:42.109355 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:26:55 crc kubenswrapper[4727]: I0929 11:26:55.107836 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:26:55 crc kubenswrapper[4727]: E0929 11:26:55.108708 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:27:07 crc kubenswrapper[4727]: I0929 11:27:07.111638 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:27:07 crc kubenswrapper[4727]: E0929 11:27:07.112800 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:27:17 crc kubenswrapper[4727]: I0929 11:27:17.434676 4727 scope.go:117] "RemoveContainer" containerID="82e40f9791267d39a35763c93a41de667a721f0089bd428df3a540f2011ae3f8" Sep 29 11:27:19 crc kubenswrapper[4727]: I0929 11:27:19.108601 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:27:19 crc kubenswrapper[4727]: E0929 11:27:19.109165 4727 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w25jt_openshift-machine-config-operator(791106af-2f47-4c80-9f20-e0bc4131f833)\"" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" Sep 29 11:27:34 crc kubenswrapper[4727]: I0929 11:27:34.108286 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:27:34 crc kubenswrapper[4727]: I0929 11:27:34.862609 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"b5484ca9e80b567378a9fcd10a8b6880f95ffbced14ec35842c967fb796e8484"} Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.407671 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bc65c799d-pbjqp_7529e49e-8904-4b11-ba31-ed805fc0fee2/barbican-api/0.log" Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.411385 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bc65c799d-pbjqp_7529e49e-8904-4b11-ba31-ed805fc0fee2/barbican-api-log/0.log" Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.652289 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b58bf9d-2f9w7_978ca6db-0819-4482-a407-a11ea6ebc3f2/barbican-keystone-listener/0.log" Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.684773 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b58bf9d-2f9w7_978ca6db-0819-4482-a407-a11ea6ebc3f2/barbican-keystone-listener-log/0.log" Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.874215 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb4b7877c-kksr7_134d320a-33b6-4df1-bf0e-da396f28f90e/barbican-worker/0.log" Sep 29 11:27:38 crc kubenswrapper[4727]: I0929 11:27:38.990887 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb4b7877c-kksr7_134d320a-33b6-4df1-bf0e-da396f28f90e/barbican-worker-log/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.074220 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4bhhl_21709a4b-d29f-47a8-acf8-ba9c1e2415c3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.291554 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/ceilometer-notification-agent/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.338590 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/ceilometer-central-agent/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.465995 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/proxy-httpd/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.508082 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa89dc0c-297d-491c-a52d-41e14a330179/sg-core/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.696931 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0248c486-ae28-4a7c-96f6-e97cde83ae7b/cinder-api-log/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.745260 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0248c486-ae28-4a7c-96f6-e97cde83ae7b/cinder-api/0.log" Sep 29 11:27:39 crc kubenswrapper[4727]: I0929 11:27:39.961518 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ae9bf7b1-b3a8-4b34-8930-3751c0230705/probe/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.005567 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ae9bf7b1-b3a8-4b34-8930-3751c0230705/cinder-scheduler/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.190585 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zr96k_e626458d-df81-47e7-aa33-47f32d67759d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.306046 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-x5dgr_d1b76565-6e67-46cf-9cc6-0f1b70d814a6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.475966 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/init/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.697684 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/dnsmasq-dns/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.707590 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6fb577f7d7-w6xqv_9e9ba97f-39de-4e67-a66c-4c7b575f161c/init/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.896539 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ctsp7_96f1c41e-9277-4df7-afd8-711f2efabcf6/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:40 crc kubenswrapper[4727]: I0929 11:27:40.937150 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_56bbd245-faf2-4c21-b4c9-172a501374fd/glance-httpd/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.065866 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_56bbd245-faf2-4c21-b4c9-172a501374fd/glance-log/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.163525 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b1ab9d58-dea7-47fc-a259-5b9b8f7a7230/glance-httpd/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.253190 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b1ab9d58-dea7-47fc-a259-5b9b8f7a7230/glance-log/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.434703 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-54cbdfc6b8-xl7cd_4b7b8921-ae97-45b0-a082-07db49561514/horizon/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.646500 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xqw72_d3ec4e67-b454-43bd-ab50-0257dc2e517d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.815633 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-tzhl4_99127a9e-ae57-42b7-9392-e3f3d7ab43ff/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:41 crc kubenswrapper[4727]: I0929 11:27:41.852550 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-54cbdfc6b8-xl7cd_4b7b8921-ae97-45b0-a082-07db49561514/horizon-log/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.095595 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64f85d6496-qmbdm_9eae012b-b92e-44ae-8396-f7331f267787/keystone-api/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.099654 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29319061-6r9mf_de7eade9-d522-4189-8a6c-60c26f0fcec7/keystone-cron/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.252989 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_bbc0072e-f90a-4518-8505-bee6f68abbaa/kube-state-metrics/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.311692 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xskrj_5fd8b89e-ffdd-4382-9bc1-4f35d7547cea/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.659890 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68ff49cc97-zxm88_326c2e7e-51f1-4f8c-9d91-4572393d8c11/neutron-httpd/0.log" Sep 29 11:27:42 crc kubenswrapper[4727]: I0929 11:27:42.674253 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68ff49cc97-zxm88_326c2e7e-51f1-4f8c-9d91-4572393d8c11/neutron-api/0.log" Sep 29 11:27:43 crc kubenswrapper[4727]: I0929 11:27:43.063891 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-65j2t_4bd4cabe-bcac-4332-b7be-326693d4bc95/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:43 crc kubenswrapper[4727]: I0929 11:27:43.552432 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_725e53c7-0270-451e-8bde-851f9a60a8ab/nova-api-log/0.log" Sep 29 11:27:43 crc kubenswrapper[4727]: I0929 11:27:43.805885 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_45ccefb3-e0b3-4484-af4b-29b81d687507/nova-cell0-conductor-conductor/0.log" Sep 29 11:27:43 crc kubenswrapper[4727]: I0929 11:27:43.923387 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_725e53c7-0270-451e-8bde-851f9a60a8ab/nova-api-api/0.log" Sep 29 11:27:44 crc kubenswrapper[4727]: I0929 11:27:44.192738 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2522d266-4b11-4a05-96fc-69d5847a63c0/nova-cell1-conductor-conductor/0.log" Sep 29 11:27:44 crc kubenswrapper[4727]: I0929 11:27:44.288631 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6267f772-555e-4c6f-b937-6d8a4c61dab3/nova-cell1-novncproxy-novncproxy/0.log" Sep 29 11:27:44 crc kubenswrapper[4727]: I0929 11:27:44.448550 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-b78rw_a3618996-7c1c-480d-b868-25e17445d8a0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:44 crc kubenswrapper[4727]: I0929 11:27:44.709859 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f944aa92-5fe8-4100-9faf-00ef0443d80f/nova-metadata-log/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.154752 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_41ff8f24-0d46-49e0-88be-7b4d3c33d619/nova-scheduler-scheduler/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.238771 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/mysql-bootstrap/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.404891 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/mysql-bootstrap/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.434064 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ac9728f6-fdb6-47ba-a17c-4220301c2a88/galera/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.660329 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/mysql-bootstrap/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.874764 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/galera/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.963215 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2d25415e-3c8d-4cd3-ab20-83a815a5e39f/mysql-bootstrap/0.log" Sep 29 11:27:45 crc kubenswrapper[4727]: I0929 11:27:45.977683 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f944aa92-5fe8-4100-9faf-00ef0443d80f/nova-metadata-metadata/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.185657 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_60736cbf-afd3-410e-87b8-bcf852ff2221/openstackclient/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.328677 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rvsjj_ad7a1be1-e75b-47f3-8b78-48f30fa37940/openstack-network-exporter/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.472496 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server-init/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.709371 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.714600 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovs-vswitchd/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.718998 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhhrr_7cb18806-a80a-491a-8ade-6371af8b54a9/ovsdb-server-init/0.log" Sep 29 11:27:46 crc kubenswrapper[4727]: I0929 11:27:46.911096 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x86c4_36e70287-3b92-41e6-a056-fb29d1e03772/ovn-controller/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.133438 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2db0e30f-219c-4539-8f02-3f64f9f23a50/openstack-network-exporter/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.134232 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-l9gxq_1dc57566-f99c-4041-a617-199b93c4e4eb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.339288 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2db0e30f-219c-4539-8f02-3f64f9f23a50/ovn-northd/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.484065 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4bb550f2-83cb-472e-9e1e-b5da4779b42d/openstack-network-exporter/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.499254 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4bb550f2-83cb-472e-9e1e-b5da4779b42d/ovsdbserver-nb/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.687708 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13550429-27b2-4d79-8a60-447c329c77f6/openstack-network-exporter/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.720474 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13550429-27b2-4d79-8a60-447c329c77f6/ovsdbserver-sb/0.log" Sep 29 11:27:47 crc kubenswrapper[4727]: I0929 11:27:47.951613 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c44b4b898-64gpw_1b18f96d-e3b8-4758-8f71-b75a40cc9fbe/placement-api/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.063009 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c44b4b898-64gpw_1b18f96d-e3b8-4758-8f71-b75a40cc9fbe/placement-log/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.156785 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/setup-container/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.445319 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/setup-container/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.456350 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f320dfce-50ce-41ea-b18f-137cd39add94/rabbitmq/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.620429 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/setup-container/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.819640 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/setup-container/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.840438 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cb8e279-f15d-46d1-a9fe-38618ca7cc90/rabbitmq/0.log" Sep 29 11:27:48 crc kubenswrapper[4727]: I0929 11:27:48.984052 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dctzx_b6b1b4ed-2c0b-41f3-a542-e6fefaccc416/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.132522 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zbjd8_eac8b453-a168-458e-9e0b-28909a6323c4/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.285355 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2x92j_874119b7-b7ce-4e87-a254-892df476efb0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.412297 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tfmmh_bc9712d6-5bbe-4749-acc0-a252a76a0ada/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.626855 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vx2sv_28a41bdb-b145-481a-9835-e08ce95e2e17/ssh-known-hosts-edpm-deployment/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.830641 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-598db7f995-m4hgm_87ecfca9-38cc-4b74-88b7-2d56d8f5638d/proxy-server/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.911986 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-598db7f995-m4hgm_87ecfca9-38cc-4b74-88b7-2d56d8f5638d/proxy-httpd/0.log" Sep 29 11:27:49 crc kubenswrapper[4727]: I0929 11:27:49.994771 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6jpg2_83f5e1ea-5026-48dc-a5a7-d6d841b45376/swift-ring-rebalance/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.127958 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-auditor/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.255966 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-reaper/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.394047 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-replicator/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.417106 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/account-server/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.495991 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-auditor/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.608593 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-server/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.679637 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-replicator/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.727700 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/container-updater/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.807563 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-auditor/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.932059 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-expirer/0.log" Sep 29 11:27:50 crc kubenswrapper[4727]: I0929 11:27:50.971389 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-replicator/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.054101 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-server/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.136609 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/object-updater/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.177525 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/rsync/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.244736 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7a6a9dee-ca84-4c2d-abc9-613bccd90764/swift-recon-cron/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.418635 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-r6bvw_77fc39d5-f092-4f06-9a6b-5d156935de57/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.638514 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0e997369-3d49-4bc4-95ec-77c503cda426/tempest-tests-tempest-tests-runner/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.731988 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_0518c511-9f69-42dd-88e3-661f13f8f019/test-operator-logs-container/0.log" Sep 29 11:27:51 crc kubenswrapper[4727]: I0929 11:27:51.986715 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tdr8c_a2a42153-5fde-492b-9564-fded49a50ebd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 29 11:28:01 crc kubenswrapper[4727]: I0929 11:28:01.480612 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8f2d7f61-fe27-4af2-8ef4-05d247fde09d/memcached/0.log" Sep 29 11:28:31 crc kubenswrapper[4727]: I0929 11:28:31.360052 4727 generic.go:334] "Generic (PLEG): container finished" podID="79a50198-2083-42d0-bde4-3879b2fb4ec6" containerID="40940ba9d8707d663eb3a397c737eb8a71728dbbbb910be5da2350828e4a9b5a" exitCode=0 Sep 29 11:28:31 crc kubenswrapper[4727]: I0929 11:28:31.360161 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" event={"ID":"79a50198-2083-42d0-bde4-3879b2fb4ec6","Type":"ContainerDied","Data":"40940ba9d8707d663eb3a397c737eb8a71728dbbbb910be5da2350828e4a9b5a"} Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.503918 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.534467 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-nmtn7"] Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.542017 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-nmtn7"] Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.680955 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvzhd\" (UniqueName: \"kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd\") pod \"79a50198-2083-42d0-bde4-3879b2fb4ec6\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.681057 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host\") pod \"79a50198-2083-42d0-bde4-3879b2fb4ec6\" (UID: \"79a50198-2083-42d0-bde4-3879b2fb4ec6\") " Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.681166 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host" (OuterVolumeSpecName: "host") pod "79a50198-2083-42d0-bde4-3879b2fb4ec6" (UID: "79a50198-2083-42d0-bde4-3879b2fb4ec6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.681550 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/79a50198-2083-42d0-bde4-3879b2fb4ec6-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.686955 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd" (OuterVolumeSpecName: "kube-api-access-zvzhd") pod "79a50198-2083-42d0-bde4-3879b2fb4ec6" (UID: "79a50198-2083-42d0-bde4-3879b2fb4ec6"). InnerVolumeSpecName "kube-api-access-zvzhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:28:32 crc kubenswrapper[4727]: I0929 11:28:32.783240 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvzhd\" (UniqueName: \"kubernetes.io/projected/79a50198-2083-42d0-bde4-3879b2fb4ec6-kube-api-access-zvzhd\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.119985 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a50198-2083-42d0-bde4-3879b2fb4ec6" path="/var/lib/kubelet/pods/79a50198-2083-42d0-bde4-3879b2fb4ec6/volumes" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.379242 4727 scope.go:117] "RemoveContainer" containerID="40940ba9d8707d663eb3a397c737eb8a71728dbbbb910be5da2350828e4a9b5a" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.379353 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-nmtn7" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.723328 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-2rc57"] Sep 29 11:28:33 crc kubenswrapper[4727]: E0929 11:28:33.723735 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a50198-2083-42d0-bde4-3879b2fb4ec6" containerName="container-00" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.723747 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a50198-2083-42d0-bde4-3879b2fb4ec6" containerName="container-00" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.723931 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a50198-2083-42d0-bde4-3879b2fb4ec6" containerName="container-00" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.724602 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.912701 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:33 crc kubenswrapper[4727]: I0929 11:28:33.912835 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjhw5\" (UniqueName: \"kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.014256 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjhw5\" (UniqueName: \"kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.014403 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.014576 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.033020 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjhw5\" (UniqueName: \"kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5\") pod \"crc-debug-2rc57\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.043582 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:34 crc kubenswrapper[4727]: W0929 11:28:34.072208 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4651308c_9373_4593_9edb_0b59e3152552.slice/crio-84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a WatchSource:0}: Error finding container 84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a: Status 404 returned error can't find the container with id 84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.390288 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" event={"ID":"4651308c-9373-4593-9edb-0b59e3152552","Type":"ContainerStarted","Data":"6f694587e4d0b0b63e332cae54d6a17434484ccc6622349904ae710c2baba388"} Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.390702 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" event={"ID":"4651308c-9373-4593-9edb-0b59e3152552","Type":"ContainerStarted","Data":"84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a"} Sep 29 11:28:34 crc kubenswrapper[4727]: I0929 11:28:34.406666 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" podStartSLOduration=1.406646586 podStartE2EDuration="1.406646586s" podCreationTimestamp="2025-09-29 11:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 11:28:34.402982754 +0000 UTC m=+3984.576296116" watchObservedRunningTime="2025-09-29 11:28:34.406646586 +0000 UTC m=+3984.579959948" Sep 29 11:28:35 crc kubenswrapper[4727]: I0929 11:28:35.401458 4727 generic.go:334] "Generic (PLEG): container finished" podID="4651308c-9373-4593-9edb-0b59e3152552" containerID="6f694587e4d0b0b63e332cae54d6a17434484ccc6622349904ae710c2baba388" exitCode=0 Sep 29 11:28:35 crc kubenswrapper[4727]: I0929 11:28:35.401545 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" event={"ID":"4651308c-9373-4593-9edb-0b59e3152552","Type":"ContainerDied","Data":"6f694587e4d0b0b63e332cae54d6a17434484ccc6622349904ae710c2baba388"} Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.521060 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.552459 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjhw5\" (UniqueName: \"kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5\") pod \"4651308c-9373-4593-9edb-0b59e3152552\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.552675 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host\") pod \"4651308c-9373-4593-9edb-0b59e3152552\" (UID: \"4651308c-9373-4593-9edb-0b59e3152552\") " Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.552804 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host" (OuterVolumeSpecName: "host") pod "4651308c-9373-4593-9edb-0b59e3152552" (UID: "4651308c-9373-4593-9edb-0b59e3152552"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.553390 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4651308c-9373-4593-9edb-0b59e3152552-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.561073 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5" (OuterVolumeSpecName: "kube-api-access-sjhw5") pod "4651308c-9373-4593-9edb-0b59e3152552" (UID: "4651308c-9373-4593-9edb-0b59e3152552"). InnerVolumeSpecName "kube-api-access-sjhw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:28:36 crc kubenswrapper[4727]: I0929 11:28:36.654812 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjhw5\" (UniqueName: \"kubernetes.io/projected/4651308c-9373-4593-9edb-0b59e3152552-kube-api-access-sjhw5\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:37 crc kubenswrapper[4727]: I0929 11:28:37.421863 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" event={"ID":"4651308c-9373-4593-9edb-0b59e3152552","Type":"ContainerDied","Data":"84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a"} Sep 29 11:28:37 crc kubenswrapper[4727]: I0929 11:28:37.421937 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84a2fa0c9cc7154b1b4b582314737a0e60cf4d6c5923211ee0dabd58ea87839a" Sep 29 11:28:37 crc kubenswrapper[4727]: I0929 11:28:37.422024 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-2rc57" Sep 29 11:28:41 crc kubenswrapper[4727]: I0929 11:28:41.406552 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-2rc57"] Sep 29 11:28:41 crc kubenswrapper[4727]: I0929 11:28:41.413982 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-2rc57"] Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.573132 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-x9zxj"] Sep 29 11:28:42 crc kubenswrapper[4727]: E0929 11:28:42.574103 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4651308c-9373-4593-9edb-0b59e3152552" containerName="container-00" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.574125 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="4651308c-9373-4593-9edb-0b59e3152552" containerName="container-00" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.574390 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="4651308c-9373-4593-9edb-0b59e3152552" containerName="container-00" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.575215 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.742590 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.742665 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqcsk\" (UniqueName: \"kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.844383 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.844483 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqcsk\" (UniqueName: \"kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.844507 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.871139 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqcsk\" (UniqueName: \"kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk\") pod \"crc-debug-x9zxj\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:42 crc kubenswrapper[4727]: I0929 11:28:42.896396 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.128316 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4651308c-9373-4593-9edb-0b59e3152552" path="/var/lib/kubelet/pods/4651308c-9373-4593-9edb-0b59e3152552/volumes" Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.481941 4727 generic.go:334] "Generic (PLEG): container finished" podID="52d35add-e5e5-4556-84f7-effa25051ec6" containerID="b5009078a1da88f2cc4c699f701f8992c9fb93a717930c348182d9437e77b349" exitCode=0 Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.482045 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" event={"ID":"52d35add-e5e5-4556-84f7-effa25051ec6","Type":"ContainerDied","Data":"b5009078a1da88f2cc4c699f701f8992c9fb93a717930c348182d9437e77b349"} Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.482689 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" event={"ID":"52d35add-e5e5-4556-84f7-effa25051ec6","Type":"ContainerStarted","Data":"b8111e7332e07f6bf7169e318504edfadee95f6ed2649ff82f4d29ce3a66f75b"} Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.526209 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-x9zxj"] Sep 29 11:28:43 crc kubenswrapper[4727]: I0929 11:28:43.536578 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hfjd5/crc-debug-x9zxj"] Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.595834 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.779482 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqcsk\" (UniqueName: \"kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk\") pod \"52d35add-e5e5-4556-84f7-effa25051ec6\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.779968 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host\") pod \"52d35add-e5e5-4556-84f7-effa25051ec6\" (UID: \"52d35add-e5e5-4556-84f7-effa25051ec6\") " Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.780307 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host" (OuterVolumeSpecName: "host") pod "52d35add-e5e5-4556-84f7-effa25051ec6" (UID: "52d35add-e5e5-4556-84f7-effa25051ec6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.780473 4727 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/52d35add-e5e5-4556-84f7-effa25051ec6-host\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.790188 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk" (OuterVolumeSpecName: "kube-api-access-vqcsk") pod "52d35add-e5e5-4556-84f7-effa25051ec6" (UID: "52d35add-e5e5-4556-84f7-effa25051ec6"). InnerVolumeSpecName "kube-api-access-vqcsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:28:44 crc kubenswrapper[4727]: I0929 11:28:44.881726 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqcsk\" (UniqueName: \"kubernetes.io/projected/52d35add-e5e5-4556-84f7-effa25051ec6-kube-api-access-vqcsk\") on node \"crc\" DevicePath \"\"" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.121477 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52d35add-e5e5-4556-84f7-effa25051ec6" path="/var/lib/kubelet/pods/52d35add-e5e5-4556-84f7-effa25051ec6/volumes" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.500632 4727 scope.go:117] "RemoveContainer" containerID="b5009078a1da88f2cc4c699f701f8992c9fb93a717930c348182d9437e77b349" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.500673 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/crc-debug-x9zxj" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.673146 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.833493 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.841245 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:28:45 crc kubenswrapper[4727]: I0929 11:28:45.846434 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.034254 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/util/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.056253 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/pull/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.072484 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3eb1cf27ab68f4d7b7a4c33f574e98f35c88042df113ea283ffa8ccce8dvgd8_9968f970-53fe-440d-9e00-2f3409f8201c/extract/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.218585 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6495d75b5-rbtdb_44a5b923-82b5-4af1-ad00-ba65998598ff/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.301541 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-748c574d75-8s5z5_f6f41603-6f22-4086-be87-c1a0062a691d/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.307129 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6495d75b5-rbtdb_44a5b923-82b5-4af1-ad00-ba65998598ff/manager/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.439125 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-748c574d75-8s5z5_f6f41603-6f22-4086-be87-c1a0062a691d/manager/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.488993 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d74f4d695-8txnk_2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.498579 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d74f4d695-8txnk_2b0b1c44-eed5-4b52-8bde-3e3c3f1b1609/manager/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.656906 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67b5d44b7f-vx2tw_2cb520ac-7d2f-46ab-b8bc-5f515485d278/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.744086 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67b5d44b7f-vx2tw_2cb520ac-7d2f-46ab-b8bc-5f515485d278/manager/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.804279 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8ff95898-6b6gt_aef52f91-e131-49a0-84c8-ba76db16d118/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.859965 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8ff95898-6b6gt_aef52f91-e131-49a0-84c8-ba76db16d118/manager/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.921265 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-695847bc78-fcs2w_66342226-bc85-46a0-8d84-0627c26b196c/kube-rbac-proxy/0.log" Sep 29 11:28:46 crc kubenswrapper[4727]: I0929 11:28:46.988475 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-695847bc78-fcs2w_66342226-bc85-46a0-8d84-0627c26b196c/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.055406 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858cd69f49-swh7z_e5545a7b-7c61-4e6b-92e9-3a22045a2133/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.246323 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858cd69f49-swh7z_e5545a7b-7c61-4e6b-92e9-3a22045a2133/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.259643 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9fc8d5567-h5bnt_f4fac1bc-f504-4b82-8371-1c7afd6f76c7/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.267638 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9fc8d5567-h5bnt_f4fac1bc-f504-4b82-8371-1c7afd6f76c7/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.436310 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7bf498966c-vktwk_6f3ebbd5-f00d-4db3-8160-03e9dce230e4/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.514154 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7bf498966c-vktwk_6f3ebbd5-f00d-4db3-8160-03e9dce230e4/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.622742 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56cf9c6b99-5xzsf_51327d2c-111c-4201-9f5d-ec9bf7f1d31f/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.623370 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56cf9c6b99-5xzsf_51327d2c-111c-4201-9f5d-ec9bf7f1d31f/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.675887 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-687b9cf756-z5zqv_f0ad5209-9d60-488c-9cb7-3252bb844305/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.836582 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-687b9cf756-z5zqv_f0ad5209-9d60-488c-9cb7-3252bb844305/manager/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.868619 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54d766c9f9-jxvz6_ff90604b-0b99-4c43-b38a-d467d74a6376/kube-rbac-proxy/0.log" Sep 29 11:28:47 crc kubenswrapper[4727]: I0929 11:28:47.936466 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54d766c9f9-jxvz6_ff90604b-0b99-4c43-b38a-d467d74a6376/manager/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.026263 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-whv4l_3164f156-e48b-4eaa-8e29-ecf414254d3a/kube-rbac-proxy/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.111886 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-whv4l_3164f156-e48b-4eaa-8e29-ecf414254d3a/manager/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.257178 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-47bvr_495ff530-82ab-48b5-9c58-1f03614f5b5a/kube-rbac-proxy/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.269696 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-47bvr_495ff530-82ab-48b5-9c58-1f03614f5b5a/manager/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.405180 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-sp6px_a6dfa756-7117-4d5c-830b-a3dc23d86682/kube-rbac-proxy/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.460174 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-sp6px_a6dfa756-7117-4d5c-830b-a3dc23d86682/manager/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.571011 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-578979c99-jgw9d_ca7d374b-4f00-4831-b2b1-39fbf673b820/kube-rbac-proxy/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.741607 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bb87788c-jvd62_b2fb3066-6094-4799-b68d-67ab396b160e/kube-rbac-proxy/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.897224 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bb87788c-jvd62_b2fb3066-6094-4799-b68d-67ab396b160e/operator/0.log" Sep 29 11:28:48 crc kubenswrapper[4727]: I0929 11:28:48.946610 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ps6m_9d7c10ba-9139-47d6-a681-507681066bff/registry-server/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.120454 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5f95c46c78-6rsg9_19c9fecd-d2f6-453c-be12-ede6f08fcf0f/kube-rbac-proxy/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.234937 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5f95c46c78-6rsg9_19c9fecd-d2f6-453c-be12-ede6f08fcf0f/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.260873 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-774b97b48-j7hz6_9d628165-af26-41a6-b05c-e633673213ab/kube-rbac-proxy/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.360195 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-774b97b48-j7hz6_9d628165-af26-41a6-b05c-e633673213ab/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.443968 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-dgghl_120b089a-5999-450b-acd7-5f48a5a105bd/operator/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.675538 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-dsg68_7c7f0180-cbee-4a25-b069-3b11a90b2d08/kube-rbac-proxy/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.721995 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5bf96cfbc4-zstln_a64b04d3-a9a3-4c5e-a356-045f7b13c572/kube-rbac-proxy/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.751350 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-dsg68_7c7f0180-cbee-4a25-b069-3b11a90b2d08/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.785349 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-578979c99-jgw9d_ca7d374b-4f00-4831-b2b1-39fbf673b820/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.876727 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5bf96cfbc4-zstln_a64b04d3-a9a3-4c5e-a356-045f7b13c572/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.942701 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6p49k_baad2063-56c9-414d-b8da-5969bf4f6220/manager/0.log" Sep 29 11:28:49 crc kubenswrapper[4727]: I0929 11:28:49.988205 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-6p49k_baad2063-56c9-414d-b8da-5969bf4f6220/kube-rbac-proxy/0.log" Sep 29 11:28:50 crc kubenswrapper[4727]: I0929 11:28:50.065883 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-djbkf_fadeace3-e8ba-4e11-a2f7-6dee11f875df/kube-rbac-proxy/0.log" Sep 29 11:28:50 crc kubenswrapper[4727]: I0929 11:28:50.115459 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-djbkf_fadeace3-e8ba-4e11-a2f7-6dee11f875df/manager/0.log" Sep 29 11:29:05 crc kubenswrapper[4727]: I0929 11:29:05.860311 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-f7hn5_46124f8a-3360-4dfe-b468-5addce47a614/control-plane-machine-set-operator/0.log" Sep 29 11:29:06 crc kubenswrapper[4727]: I0929 11:29:06.021485 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lw586_6404327b-095b-435a-b953-b1c3b7a6f332/kube-rbac-proxy/0.log" Sep 29 11:29:06 crc kubenswrapper[4727]: I0929 11:29:06.037889 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lw586_6404327b-095b-435a-b953-b1c3b7a6f332/machine-api-operator/0.log" Sep 29 11:29:17 crc kubenswrapper[4727]: I0929 11:29:17.259191 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wxxch_aec4835b-f46f-4b6a-aab2-37154553a6aa/cert-manager-controller/0.log" Sep 29 11:29:17 crc kubenswrapper[4727]: I0929 11:29:17.435533 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-62b6k_15e48513-d028-479d-9796-ab2ecbe72f58/cert-manager-cainjector/0.log" Sep 29 11:29:17 crc kubenswrapper[4727]: I0929 11:29:17.500514 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-6w54j_4e4ebfd5-7776-4705-943f-b6ac0e6a6238/cert-manager-webhook/0.log" Sep 29 11:29:17 crc kubenswrapper[4727]: I0929 11:29:17.524885 4727 scope.go:117] "RemoveContainer" containerID="d82b3e06770a823dfed883fc8a0a412e53bd8d1f81ec6e21554c280db9e52cc8" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.088470 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-l7xjv_8a2bcf35-f31c-44b3-8a73-c6c6e2c5c2c6/nmstate-console-plugin/0.log" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.280682 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sh4hc_8109dd98-72c1-4542-a34c-1f5d911b5fd4/nmstate-handler/0.log" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.356958 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2hx48_a713c14e-bce4-44a3-940d-0a987cc6fa5b/kube-rbac-proxy/0.log" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.389918 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2hx48_a713c14e-bce4-44a3-940d-0a987cc6fa5b/nmstate-metrics/0.log" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.501183 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-5vcth_2ef4df09-2078-4ef8-9c7b-8e637b20cfcf/nmstate-operator/0.log" Sep 29 11:29:28 crc kubenswrapper[4727]: I0929 11:29:28.560665 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-shptq_7655e999-03ae-45f0-a123-1efa69626f7b/nmstate-webhook/0.log" Sep 29 11:29:43 crc kubenswrapper[4727]: I0929 11:29:43.390366 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-b8fqx_5a5b2bf7-f24e-494d-a0fc-11c6f90737bd/kube-rbac-proxy/0.log" Sep 29 11:29:43 crc kubenswrapper[4727]: I0929 11:29:43.522322 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-b8fqx_5a5b2bf7-f24e-494d-a0fc-11c6f90737bd/controller/0.log" Sep 29 11:29:43 crc kubenswrapper[4727]: I0929 11:29:43.765239 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:29:43 crc kubenswrapper[4727]: I0929 11:29:43.952595 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:29:43 crc kubenswrapper[4727]: I0929 11:29:43.990933 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.032286 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.072759 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.273242 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.290030 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.296030 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.319450 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.972807 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-metrics/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.972824 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-frr-files/0.log" Sep 29 11:29:44 crc kubenswrapper[4727]: I0929 11:29:44.985911 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/controller/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.021940 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/cp-reloader/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.184057 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/frr-metrics/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.212487 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/kube-rbac-proxy/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.255853 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/kube-rbac-proxy-frr/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.427507 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/reloader/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.519834 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-6pmt2_424bbcbe-8ea1-491c-b1b0-55285bff680e/frr-k8s-webhook-server/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.698120 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-86747676bf-9fmjz_9e45a861-9d6a-43af-9235-33e4cea43955/manager/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.895232 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b8c57d74f-nnlj8_5c0a38f9-cb25-4466-9d31-4d6bf3943989/webhook-server/0.log" Sep 29 11:29:45 crc kubenswrapper[4727]: I0929 11:29:45.957991 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vh9j9_fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e/kube-rbac-proxy/0.log" Sep 29 11:29:46 crc kubenswrapper[4727]: I0929 11:29:46.573769 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jrlvl_fb53b9ce-f888-4b6b-a3dc-be4b65aa65c2/frr/0.log" Sep 29 11:29:46 crc kubenswrapper[4727]: I0929 11:29:46.584060 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vh9j9_fa619b15-1b04-4ad9-bf3d-abfd4aee5b2e/speaker/0.log" Sep 29 11:29:49 crc kubenswrapper[4727]: I0929 11:29:49.246459 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:29:49 crc kubenswrapper[4727]: I0929 11:29:49.246779 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:29:58 crc kubenswrapper[4727]: I0929 11:29:58.651718 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:29:58 crc kubenswrapper[4727]: I0929 11:29:58.917422 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:29:58 crc kubenswrapper[4727]: I0929 11:29:58.939071 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.001520 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.153691 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/util/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.185636 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/pull/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.191557 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcpqvzk_2ae491e2-d46d-44be-9b32-877cc62c18e4/extract/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.341248 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.516276 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.549405 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.556530 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.726764 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-content/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.783978 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/extract-utilities/0.log" Sep 29 11:29:59 crc kubenswrapper[4727]: I0929 11:29:59.953890 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.161586 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29"] Sep 29 11:30:00 crc kubenswrapper[4727]: E0929 11:30:00.162087 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d35add-e5e5-4556-84f7-effa25051ec6" containerName="container-00" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.162107 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d35add-e5e5-4556-84f7-effa25051ec6" containerName="container-00" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.162387 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d35add-e5e5-4556-84f7-effa25051ec6" containerName="container-00" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.163182 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.165551 4727 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.170836 4727 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.178747 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29"] Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.239956 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.240750 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.262310 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.323898 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.324067 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.324106 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pthng\" (UniqueName: \"kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.371934 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bhmq_3410083e-6ab2-4674-8544-9ed975c2db00/registry-server/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.425388 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.425488 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pthng\" (UniqueName: \"kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.425581 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.426692 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.432015 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.443195 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pthng\" (UniqueName: \"kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng\") pod \"collect-profiles-29319090-tlk29\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.489666 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-content/0.log" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.491121 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:00 crc kubenswrapper[4727]: I0929 11:30:00.579477 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:00.749885 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:00.991681 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.205975 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lfbq8_eddf2895-0cde-4925-8ddb-98bdd9fd4f64/registry-server/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.228584 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.268839 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.481809 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/pull/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.490900 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/extract/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.521768 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96c2tst_5d62033f-fb65-4d2f-b853-d241790d6bca/util/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.681859 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nsbj_a9c2af09-6645-4cef-bba5-59c2b5b278d2/marketplace-operator/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.715633 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.950816 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.976030 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:01.997104 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.192723 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.268148 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/extract-content/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.300285 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hp9xv_6e4c7812-66fe-4e23-bb8e-91b2719840f5/registry-server/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.425643 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.673187 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.693567 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.703536 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.781384 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29"] Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.890929 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-utilities/0.log" Sep 29 11:30:02 crc kubenswrapper[4727]: I0929 11:30:02.939574 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/extract-content/0.log" Sep 29 11:30:03 crc kubenswrapper[4727]: I0929 11:30:03.133723 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9s47w_b161c1a4-0830-45fa-943b-579070b738ff/registry-server/0.log" Sep 29 11:30:03 crc kubenswrapper[4727]: I0929 11:30:03.249767 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" event={"ID":"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae","Type":"ContainerStarted","Data":"cecf67acbfc4d66659ccb27d2fc837689a377fb86e29a8312f3a485b6fc9cdfc"} Sep 29 11:30:03 crc kubenswrapper[4727]: I0929 11:30:03.249806 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" event={"ID":"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae","Type":"ContainerStarted","Data":"3711f7693453c9188634ae73a355a8f12b34d6c026f8d61b4585ac5192572a5f"} Sep 29 11:30:04 crc kubenswrapper[4727]: I0929 11:30:04.261455 4727 generic.go:334] "Generic (PLEG): container finished" podID="2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" containerID="cecf67acbfc4d66659ccb27d2fc837689a377fb86e29a8312f3a485b6fc9cdfc" exitCode=0 Sep 29 11:30:04 crc kubenswrapper[4727]: I0929 11:30:04.261532 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" event={"ID":"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae","Type":"ContainerDied","Data":"cecf67acbfc4d66659ccb27d2fc837689a377fb86e29a8312f3a485b6fc9cdfc"} Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.723368 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.835258 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume\") pod \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.835656 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume\") pod \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.835905 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pthng\" (UniqueName: \"kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng\") pod \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\" (UID: \"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae\") " Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.837197 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume" (OuterVolumeSpecName: "config-volume") pod "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" (UID: "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.842256 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng" (OuterVolumeSpecName: "kube-api-access-pthng") pod "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" (UID: "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae"). InnerVolumeSpecName "kube-api-access-pthng". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.842782 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" (UID: "2e2fc3e4-efa4-4c94-b7d4-686488ba6dae"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.937445 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pthng\" (UniqueName: \"kubernetes.io/projected/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-kube-api-access-pthng\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.937625 4727 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:05 crc kubenswrapper[4727]: I0929 11:30:05.937677 4727 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e2fc3e4-efa4-4c94-b7d4-686488ba6dae-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:06 crc kubenswrapper[4727]: I0929 11:30:06.281413 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" event={"ID":"2e2fc3e4-efa4-4c94-b7d4-686488ba6dae","Type":"ContainerDied","Data":"3711f7693453c9188634ae73a355a8f12b34d6c026f8d61b4585ac5192572a5f"} Sep 29 11:30:06 crc kubenswrapper[4727]: I0929 11:30:06.281463 4727 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3711f7693453c9188634ae73a355a8f12b34d6c026f8d61b4585ac5192572a5f" Sep 29 11:30:06 crc kubenswrapper[4727]: I0929 11:30:06.281689 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319090-tlk29" Sep 29 11:30:06 crc kubenswrapper[4727]: I0929 11:30:06.810709 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w"] Sep 29 11:30:06 crc kubenswrapper[4727]: I0929 11:30:06.818206 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319045-zzl2w"] Sep 29 11:30:07 crc kubenswrapper[4727]: I0929 11:30:07.118321 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d1abf6-0c7c-422a-90c9-2e5c9b888061" path="/var/lib/kubelet/pods/61d1abf6-0c7c-422a-90c9-2e5c9b888061/volumes" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.773530 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:13 crc kubenswrapper[4727]: E0929 11:30:13.774412 4727 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" containerName="collect-profiles" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.774424 4727 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" containerName="collect-profiles" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.774618 4727 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2fc3e4-efa4-4c94-b7d4-686488ba6dae" containerName="collect-profiles" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.775962 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.801101 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.888123 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfrd6\" (UniqueName: \"kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.888459 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.888502 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.990523 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfrd6\" (UniqueName: \"kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.990570 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.990606 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.991013 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:13 crc kubenswrapper[4727]: I0929 11:30:13.991126 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.014315 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfrd6\" (UniqueName: \"kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6\") pod \"redhat-operators-48tz6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.093121 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.587579 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:14 crc kubenswrapper[4727]: W0929 11:30:14.589902 4727 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda87ab7a6_4e33_4bcd_930b_6b6d67a9c6b6.slice/crio-5e3c18dfcb59dfeb0f9f90a5b3ec5334853d3521ade6878946656bea5a0fa192 WatchSource:0}: Error finding container 5e3c18dfcb59dfeb0f9f90a5b3ec5334853d3521ade6878946656bea5a0fa192: Status 404 returned error can't find the container with id 5e3c18dfcb59dfeb0f9f90a5b3ec5334853d3521ade6878946656bea5a0fa192 Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.774921 4727 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.777165 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.787981 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.935427 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.936141 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsrp9\" (UniqueName: \"kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:14 crc kubenswrapper[4727]: I0929 11:30:14.936882 4727 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.038941 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.039014 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.039050 4727 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsrp9\" (UniqueName: \"kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.039535 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.039590 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.304436 4727 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsrp9\" (UniqueName: \"kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9\") pod \"certified-operators-kfmb9\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.364055 4727 generic.go:334] "Generic (PLEG): container finished" podID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" containerID="3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6" exitCode=0 Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.364166 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerDied","Data":"3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6"} Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.364418 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerStarted","Data":"5e3c18dfcb59dfeb0f9f90a5b3ec5334853d3521ade6878946656bea5a0fa192"} Sep 29 11:30:15 crc kubenswrapper[4727]: I0929 11:30:15.436925 4727 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:16 crc kubenswrapper[4727]: I0929 11:30:16.023480 4727 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:30:16 crc kubenswrapper[4727]: I0929 11:30:16.378788 4727 generic.go:334] "Generic (PLEG): container finished" podID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerID="2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c" exitCode=0 Sep 29 11:30:16 crc kubenswrapper[4727]: I0929 11:30:16.379081 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerDied","Data":"2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c"} Sep 29 11:30:16 crc kubenswrapper[4727]: I0929 11:30:16.379111 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerStarted","Data":"f94482be97713f19685a69c3db9c9cc4e76f473e4c8ba5ce9ba20aa0be179168"} Sep 29 11:30:17 crc kubenswrapper[4727]: I0929 11:30:17.388918 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerStarted","Data":"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62"} Sep 29 11:30:17 crc kubenswrapper[4727]: I0929 11:30:17.636936 4727 scope.go:117] "RemoveContainer" containerID="f849a508f8c6124d0eaba17ac34eed0c93600bcd61795235abd655cb3ef700c9" Sep 29 11:30:18 crc kubenswrapper[4727]: I0929 11:30:18.399065 4727 generic.go:334] "Generic (PLEG): container finished" podID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" containerID="1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62" exitCode=0 Sep 29 11:30:18 crc kubenswrapper[4727]: I0929 11:30:18.399112 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerDied","Data":"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62"} Sep 29 11:30:19 crc kubenswrapper[4727]: I0929 11:30:19.247065 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:30:19 crc kubenswrapper[4727]: I0929 11:30:19.247525 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:30:19 crc kubenswrapper[4727]: I0929 11:30:19.414245 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerStarted","Data":"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2"} Sep 29 11:30:20 crc kubenswrapper[4727]: I0929 11:30:20.426376 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerStarted","Data":"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66"} Sep 29 11:30:20 crc kubenswrapper[4727]: I0929 11:30:20.448269 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48tz6" podStartSLOduration=3.450532194 podStartE2EDuration="7.448244658s" podCreationTimestamp="2025-09-29 11:30:13 +0000 UTC" firstStartedPulling="2025-09-29 11:30:15.36603172 +0000 UTC m=+4085.539345082" lastFinishedPulling="2025-09-29 11:30:19.363744184 +0000 UTC m=+4089.537057546" observedRunningTime="2025-09-29 11:30:20.444147435 +0000 UTC m=+4090.617460797" watchObservedRunningTime="2025-09-29 11:30:20.448244658 +0000 UTC m=+4090.621558020" Sep 29 11:30:24 crc kubenswrapper[4727]: I0929 11:30:24.094064 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:24 crc kubenswrapper[4727]: I0929 11:30:24.096877 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:25 crc kubenswrapper[4727]: I0929 11:30:25.151771 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48tz6" podUID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" containerName="registry-server" probeResult="failure" output=< Sep 29 11:30:25 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 11:30:25 crc kubenswrapper[4727]: > Sep 29 11:30:31 crc kubenswrapper[4727]: I0929 11:30:31.549209 4727 generic.go:334] "Generic (PLEG): container finished" podID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerID="6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2" exitCode=0 Sep 29 11:30:31 crc kubenswrapper[4727]: I0929 11:30:31.549297 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerDied","Data":"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2"} Sep 29 11:30:33 crc kubenswrapper[4727]: I0929 11:30:33.574474 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerStarted","Data":"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98"} Sep 29 11:30:33 crc kubenswrapper[4727]: I0929 11:30:33.593261 4727 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfmb9" podStartSLOduration=3.23411197 podStartE2EDuration="19.59324565s" podCreationTimestamp="2025-09-29 11:30:14 +0000 UTC" firstStartedPulling="2025-09-29 11:30:16.468214157 +0000 UTC m=+4086.641527519" lastFinishedPulling="2025-09-29 11:30:32.827347837 +0000 UTC m=+4103.000661199" observedRunningTime="2025-09-29 11:30:33.592462071 +0000 UTC m=+4103.765775443" watchObservedRunningTime="2025-09-29 11:30:33.59324565 +0000 UTC m=+4103.766559012" Sep 29 11:30:34 crc kubenswrapper[4727]: I0929 11:30:34.173545 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:34 crc kubenswrapper[4727]: I0929 11:30:34.248904 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:34 crc kubenswrapper[4727]: I0929 11:30:34.810584 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:35 crc kubenswrapper[4727]: I0929 11:30:35.437802 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:35 crc kubenswrapper[4727]: I0929 11:30:35.437896 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:30:35 crc kubenswrapper[4727]: I0929 11:30:35.591502 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-48tz6" podUID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" containerName="registry-server" containerID="cri-o://1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66" gracePeriod=2 Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.066851 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.137983 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfrd6\" (UniqueName: \"kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6\") pod \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.138206 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities\") pod \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.138236 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content\") pod \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\" (UID: \"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6\") " Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.151941 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities" (OuterVolumeSpecName: "utilities") pod "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" (UID: "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.152592 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6" (OuterVolumeSpecName: "kube-api-access-jfrd6") pod "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" (UID: "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6"). InnerVolumeSpecName "kube-api-access-jfrd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.227386 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" (UID: "a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.242487 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.242531 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.242553 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfrd6\" (UniqueName: \"kubernetes.io/projected/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6-kube-api-access-jfrd6\") on node \"crc\" DevicePath \"\"" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.487056 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kfmb9" podUID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerName="registry-server" probeResult="failure" output=< Sep 29 11:30:36 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 11:30:36 crc kubenswrapper[4727]: > Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.608927 4727 generic.go:334] "Generic (PLEG): container finished" podID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" containerID="1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66" exitCode=0 Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.608989 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerDied","Data":"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66"} Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.609064 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48tz6" event={"ID":"a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6","Type":"ContainerDied","Data":"5e3c18dfcb59dfeb0f9f90a5b3ec5334853d3521ade6878946656bea5a0fa192"} Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.609018 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48tz6" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.609084 4727 scope.go:117] "RemoveContainer" containerID="1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.629725 4727 scope.go:117] "RemoveContainer" containerID="1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.654064 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.665068 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-48tz6"] Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.741737 4727 scope.go:117] "RemoveContainer" containerID="3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.767214 4727 scope.go:117] "RemoveContainer" containerID="1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66" Sep 29 11:30:36 crc kubenswrapper[4727]: E0929 11:30:36.769685 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66\": container with ID starting with 1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66 not found: ID does not exist" containerID="1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.769887 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66"} err="failed to get container status \"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66\": rpc error: code = NotFound desc = could not find container \"1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66\": container with ID starting with 1e5423a180f4167db5ef4c801a676b334aa6b1efd9fc689810c634d56df43d66 not found: ID does not exist" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.769920 4727 scope.go:117] "RemoveContainer" containerID="1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62" Sep 29 11:30:36 crc kubenswrapper[4727]: E0929 11:30:36.770625 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62\": container with ID starting with 1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62 not found: ID does not exist" containerID="1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.770838 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62"} err="failed to get container status \"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62\": rpc error: code = NotFound desc = could not find container \"1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62\": container with ID starting with 1bdbbd2a42fcac8d0511ff23bab737933e1edfeb90247fc1782259c5952e7a62 not found: ID does not exist" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.770873 4727 scope.go:117] "RemoveContainer" containerID="3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6" Sep 29 11:30:36 crc kubenswrapper[4727]: E0929 11:30:36.771498 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6\": container with ID starting with 3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6 not found: ID does not exist" containerID="3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6" Sep 29 11:30:36 crc kubenswrapper[4727]: I0929 11:30:36.771544 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6"} err="failed to get container status \"3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6\": rpc error: code = NotFound desc = could not find container \"3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6\": container with ID starting with 3b6fce1dfa07978eac0c30d96fd66261aae6adc182d199ea20de854ffac50ef6 not found: ID does not exist" Sep 29 11:30:37 crc kubenswrapper[4727]: I0929 11:30:37.118373 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6" path="/var/lib/kubelet/pods/a87ab7a6-4e33-4bcd-930b-6b6d67a9c6b6/volumes" Sep 29 11:30:46 crc kubenswrapper[4727]: I0929 11:30:46.490707 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kfmb9" podUID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerName="registry-server" probeResult="failure" output=< Sep 29 11:30:46 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 11:30:46 crc kubenswrapper[4727]: > Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.246184 4727 patch_prober.go:28] interesting pod/machine-config-daemon-w25jt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.246705 4727 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.246774 4727 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.248105 4727 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5484ca9e80b567378a9fcd10a8b6880f95ffbced14ec35842c967fb796e8484"} pod="openshift-machine-config-operator/machine-config-daemon-w25jt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.248228 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" podUID="791106af-2f47-4c80-9f20-e0bc4131f833" containerName="machine-config-daemon" containerID="cri-o://b5484ca9e80b567378a9fcd10a8b6880f95ffbced14ec35842c967fb796e8484" gracePeriod=600 Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.744992 4727 generic.go:334] "Generic (PLEG): container finished" podID="791106af-2f47-4c80-9f20-e0bc4131f833" containerID="b5484ca9e80b567378a9fcd10a8b6880f95ffbced14ec35842c967fb796e8484" exitCode=0 Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.745069 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerDied","Data":"b5484ca9e80b567378a9fcd10a8b6880f95ffbced14ec35842c967fb796e8484"} Sep 29 11:30:49 crc kubenswrapper[4727]: I0929 11:30:49.745321 4727 scope.go:117] "RemoveContainer" containerID="a72c5f9188f5be8aa4d51eaa5a708877b3bb60104691dc91158a52e75189e384" Sep 29 11:30:50 crc kubenswrapper[4727]: I0929 11:30:50.765986 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w25jt" event={"ID":"791106af-2f47-4c80-9f20-e0bc4131f833","Type":"ContainerStarted","Data":"39d80c70a851d2442ee954f93c2639b1656da41caa101e3a7d634702121ec029"} Sep 29 11:30:56 crc kubenswrapper[4727]: I0929 11:30:56.479805 4727 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kfmb9" podUID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerName="registry-server" probeResult="failure" output=< Sep 29 11:30:56 crc kubenswrapper[4727]: timeout: failed to connect service ":50051" within 1s Sep 29 11:30:56 crc kubenswrapper[4727]: > Sep 29 11:31:05 crc kubenswrapper[4727]: I0929 11:31:05.499900 4727 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:31:05 crc kubenswrapper[4727]: I0929 11:31:05.563262 4727 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:31:05 crc kubenswrapper[4727]: I0929 11:31:05.732187 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:31:06 crc kubenswrapper[4727]: I0929 11:31:06.924628 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfmb9" podUID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerName="registry-server" containerID="cri-o://c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98" gracePeriod=2 Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.392195 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.513428 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsrp9\" (UniqueName: \"kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9\") pod \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.513783 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities\") pod \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.513842 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content\") pod \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\" (UID: \"f1224a81-34ff-41ab-b30a-d6b83e0e8708\") " Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.514605 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities" (OuterVolumeSpecName: "utilities") pod "f1224a81-34ff-41ab-b30a-d6b83e0e8708" (UID: "f1224a81-34ff-41ab-b30a-d6b83e0e8708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.519630 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9" (OuterVolumeSpecName: "kube-api-access-hsrp9") pod "f1224a81-34ff-41ab-b30a-d6b83e0e8708" (UID: "f1224a81-34ff-41ab-b30a-d6b83e0e8708"). InnerVolumeSpecName "kube-api-access-hsrp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.581524 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1224a81-34ff-41ab-b30a-d6b83e0e8708" (UID: "f1224a81-34ff-41ab-b30a-d6b83e0e8708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.616286 4727 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.616327 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsrp9\" (UniqueName: \"kubernetes.io/projected/f1224a81-34ff-41ab-b30a-d6b83e0e8708-kube-api-access-hsrp9\") on node \"crc\" DevicePath \"\"" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.616356 4727 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1224a81-34ff-41ab-b30a-d6b83e0e8708-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.936886 4727 generic.go:334] "Generic (PLEG): container finished" podID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" containerID="c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98" exitCode=0 Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.936932 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerDied","Data":"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98"} Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.936959 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfmb9" event={"ID":"f1224a81-34ff-41ab-b30a-d6b83e0e8708","Type":"ContainerDied","Data":"f94482be97713f19685a69c3db9c9cc4e76f473e4c8ba5ce9ba20aa0be179168"} Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.936974 4727 scope.go:117] "RemoveContainer" containerID="c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.936988 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfmb9" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.975059 4727 scope.go:117] "RemoveContainer" containerID="6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2" Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.980597 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:31:07 crc kubenswrapper[4727]: I0929 11:31:07.989909 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfmb9"] Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.014977 4727 scope.go:117] "RemoveContainer" containerID="2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.047507 4727 scope.go:117] "RemoveContainer" containerID="c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98" Sep 29 11:31:08 crc kubenswrapper[4727]: E0929 11:31:08.048121 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98\": container with ID starting with c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98 not found: ID does not exist" containerID="c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.048157 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98"} err="failed to get container status \"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98\": rpc error: code = NotFound desc = could not find container \"c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98\": container with ID starting with c64d1731e4ef8a3464a726731607953cf929cbff1b6b89e9941443747e878f98 not found: ID does not exist" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.048189 4727 scope.go:117] "RemoveContainer" containerID="6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2" Sep 29 11:31:08 crc kubenswrapper[4727]: E0929 11:31:08.048619 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2\": container with ID starting with 6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2 not found: ID does not exist" containerID="6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.048645 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2"} err="failed to get container status \"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2\": rpc error: code = NotFound desc = could not find container \"6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2\": container with ID starting with 6d920428a47b9ed53c8a0dc9c99386f9e2ed119dd962b317ad396620a1554ab2 not found: ID does not exist" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.048665 4727 scope.go:117] "RemoveContainer" containerID="2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c" Sep 29 11:31:08 crc kubenswrapper[4727]: E0929 11:31:08.048933 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c\": container with ID starting with 2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c not found: ID does not exist" containerID="2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c" Sep 29 11:31:08 crc kubenswrapper[4727]: I0929 11:31:08.048966 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c"} err="failed to get container status \"2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c\": rpc error: code = NotFound desc = could not find container \"2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c\": container with ID starting with 2ead533b9eb5edc0aff13097431abb171cc44181bf6e9b70461b8523b0c5ee3c not found: ID does not exist" Sep 29 11:31:09 crc kubenswrapper[4727]: I0929 11:31:09.122703 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1224a81-34ff-41ab-b30a-d6b83e0e8708" path="/var/lib/kubelet/pods/f1224a81-34ff-41ab-b30a-d6b83e0e8708/volumes" Sep 29 11:32:09 crc kubenswrapper[4727]: I0929 11:32:09.787195 4727 generic.go:334] "Generic (PLEG): container finished" podID="79746d22-d0ae-40d7-a444-4eb4e162e75d" containerID="549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760" exitCode=0 Sep 29 11:32:09 crc kubenswrapper[4727]: I0929 11:32:09.787291 4727 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" event={"ID":"79746d22-d0ae-40d7-a444-4eb4e162e75d","Type":"ContainerDied","Data":"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760"} Sep 29 11:32:09 crc kubenswrapper[4727]: I0929 11:32:09.789242 4727 scope.go:117] "RemoveContainer" containerID="549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760" Sep 29 11:32:10 crc kubenswrapper[4727]: I0929 11:32:10.477801 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hfjd5_must-gather-ghqtn_79746d22-d0ae-40d7-a444-4eb4e162e75d/gather/0.log" Sep 29 11:32:21 crc kubenswrapper[4727]: I0929 11:32:21.948652 4727 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hfjd5/must-gather-ghqtn"] Sep 29 11:32:21 crc kubenswrapper[4727]: I0929 11:32:21.949883 4727 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" podUID="79746d22-d0ae-40d7-a444-4eb4e162e75d" containerName="copy" containerID="cri-o://aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1" gracePeriod=2 Sep 29 11:32:21 crc kubenswrapper[4727]: I0929 11:32:21.959074 4727 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hfjd5/must-gather-ghqtn"] Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.723591 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hfjd5_must-gather-ghqtn_79746d22-d0ae-40d7-a444-4eb4e162e75d/copy/0.log" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.724752 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.818282 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx4p5\" (UniqueName: \"kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5\") pod \"79746d22-d0ae-40d7-a444-4eb4e162e75d\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.819178 4727 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output\") pod \"79746d22-d0ae-40d7-a444-4eb4e162e75d\" (UID: \"79746d22-d0ae-40d7-a444-4eb4e162e75d\") " Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.826990 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5" (OuterVolumeSpecName: "kube-api-access-nx4p5") pod "79746d22-d0ae-40d7-a444-4eb4e162e75d" (UID: "79746d22-d0ae-40d7-a444-4eb4e162e75d"). InnerVolumeSpecName "kube-api-access-nx4p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.921529 4727 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx4p5\" (UniqueName: \"kubernetes.io/projected/79746d22-d0ae-40d7-a444-4eb4e162e75d-kube-api-access-nx4p5\") on node \"crc\" DevicePath \"\"" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.978785 4727 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hfjd5_must-gather-ghqtn_79746d22-d0ae-40d7-a444-4eb4e162e75d/copy/0.log" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.981251 4727 generic.go:334] "Generic (PLEG): container finished" podID="79746d22-d0ae-40d7-a444-4eb4e162e75d" containerID="aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1" exitCode=143 Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.981309 4727 scope.go:117] "RemoveContainer" containerID="aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1" Sep 29 11:32:22 crc kubenswrapper[4727]: I0929 11:32:22.981456 4727 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hfjd5/must-gather-ghqtn" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.012543 4727 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "79746d22-d0ae-40d7-a444-4eb4e162e75d" (UID: "79746d22-d0ae-40d7-a444-4eb4e162e75d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.021511 4727 scope.go:117] "RemoveContainer" containerID="549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.022757 4727 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/79746d22-d0ae-40d7-a444-4eb4e162e75d-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.102546 4727 scope.go:117] "RemoveContainer" containerID="aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1" Sep 29 11:32:23 crc kubenswrapper[4727]: E0929 11:32:23.106130 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1\": container with ID starting with aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1 not found: ID does not exist" containerID="aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.106182 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1"} err="failed to get container status \"aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1\": rpc error: code = NotFound desc = could not find container \"aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1\": container with ID starting with aafd4185890fee025d113fc39341b170ed503f560357b3469dcb6ba66496f9a1 not found: ID does not exist" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.106214 4727 scope.go:117] "RemoveContainer" containerID="549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760" Sep 29 11:32:23 crc kubenswrapper[4727]: E0929 11:32:23.106623 4727 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760\": container with ID starting with 549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760 not found: ID does not exist" containerID="549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.106649 4727 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760"} err="failed to get container status \"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760\": rpc error: code = NotFound desc = could not find container \"549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760\": container with ID starting with 549f52bc003ec3598d48142337e20bcf9d290789087e81090bf4aa262fb1f760 not found: ID does not exist" Sep 29 11:32:23 crc kubenswrapper[4727]: I0929 11:32:23.144668 4727 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79746d22-d0ae-40d7-a444-4eb4e162e75d" path="/var/lib/kubelet/pods/79746d22-d0ae-40d7-a444-4eb4e162e75d/volumes" Sep 29 11:32:48 crc kubenswrapper[4727]: I0929 11:32:48.133904 4727 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-598db7f995-m4hgm" podUID="87ecfca9-38cc-4b74-88b7-2d56d8f5638d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515066467203024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015066467204017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015066456370016521 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015066456370015471 5ustar corecore